var/home/core/zuul-output/0000755000175000017500000000000015120463412014523 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015120510640015463 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005432315515120510632017700 0ustar rootrootDec 17 08:21:03 crc systemd[1]: Starting Kubernetes Kubelet... Dec 17 08:21:03 crc restorecon[4578]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:03 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 17 08:21:04 crc restorecon[4578]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 17 08:21:04 crc kubenswrapper[4966]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 17 08:21:04 crc kubenswrapper[4966]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 17 08:21:04 crc kubenswrapper[4966]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 17 08:21:04 crc kubenswrapper[4966]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 17 08:21:04 crc kubenswrapper[4966]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 17 08:21:04 crc kubenswrapper[4966]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.574398 4966 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577605 4966 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577627 4966 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577633 4966 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577638 4966 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577642 4966 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577646 4966 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577654 4966 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577659 4966 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577664 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577668 4966 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577673 4966 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577677 4966 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577681 4966 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577689 4966 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577694 4966 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577699 4966 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577703 4966 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577707 4966 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577716 4966 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577720 4966 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577724 4966 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577729 4966 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577732 4966 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577739 4966 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577743 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577748 4966 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577753 4966 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577757 4966 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577761 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577769 4966 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577773 4966 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577776 4966 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577781 4966 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577785 4966 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577788 4966 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577792 4966 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577796 4966 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577799 4966 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577803 4966 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577806 4966 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577810 4966 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577814 4966 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577821 4966 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577824 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577828 4966 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577831 4966 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577835 4966 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577839 4966 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577843 4966 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577847 4966 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577851 4966 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577855 4966 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577860 4966 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577867 4966 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577890 4966 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577893 4966 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577897 4966 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577901 4966 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577904 4966 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577908 4966 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577911 4966 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577915 4966 feature_gate.go:330] unrecognized feature gate: Example Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577918 4966 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577922 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577926 4966 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577929 4966 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577935 4966 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577939 4966 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577943 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577947 4966 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.577951 4966 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578089 4966 flags.go:64] FLAG: --address="0.0.0.0" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578101 4966 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578109 4966 flags.go:64] FLAG: --anonymous-auth="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578115 4966 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578121 4966 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578127 4966 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578136 4966 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578142 4966 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578146 4966 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578151 4966 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578155 4966 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578159 4966 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578164 4966 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578169 4966 flags.go:64] FLAG: --cgroup-root="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578176 4966 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578180 4966 flags.go:64] FLAG: --client-ca-file="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578184 4966 flags.go:64] FLAG: --cloud-config="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578188 4966 flags.go:64] FLAG: --cloud-provider="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578192 4966 flags.go:64] FLAG: --cluster-dns="[]" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578198 4966 flags.go:64] FLAG: --cluster-domain="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578203 4966 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578207 4966 flags.go:64] FLAG: --config-dir="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578214 4966 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578219 4966 flags.go:64] FLAG: --container-log-max-files="5" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578226 4966 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578232 4966 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578238 4966 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578243 4966 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578248 4966 flags.go:64] FLAG: --contention-profiling="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578252 4966 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578256 4966 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578265 4966 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578269 4966 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578275 4966 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578283 4966 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578287 4966 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578292 4966 flags.go:64] FLAG: --enable-load-reader="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578296 4966 flags.go:64] FLAG: --enable-server="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578301 4966 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578311 4966 flags.go:64] FLAG: --event-burst="100" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578316 4966 flags.go:64] FLAG: --event-qps="50" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578326 4966 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578334 4966 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578340 4966 flags.go:64] FLAG: --eviction-hard="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578347 4966 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578352 4966 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578358 4966 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578364 4966 flags.go:64] FLAG: --eviction-soft="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578373 4966 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578380 4966 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578385 4966 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578390 4966 flags.go:64] FLAG: --experimental-mounter-path="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578395 4966 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578399 4966 flags.go:64] FLAG: --fail-swap-on="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578404 4966 flags.go:64] FLAG: --feature-gates="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578410 4966 flags.go:64] FLAG: --file-check-frequency="20s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578417 4966 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578421 4966 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578426 4966 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578430 4966 flags.go:64] FLAG: --healthz-port="10248" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578435 4966 flags.go:64] FLAG: --help="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578439 4966 flags.go:64] FLAG: --hostname-override="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578443 4966 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578466 4966 flags.go:64] FLAG: --http-check-frequency="20s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578471 4966 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578478 4966 flags.go:64] FLAG: --image-credential-provider-config="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578482 4966 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578486 4966 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578491 4966 flags.go:64] FLAG: --image-service-endpoint="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578495 4966 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578500 4966 flags.go:64] FLAG: --kube-api-burst="100" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578504 4966 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578509 4966 flags.go:64] FLAG: --kube-api-qps="50" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578515 4966 flags.go:64] FLAG: --kube-reserved="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578520 4966 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578525 4966 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578531 4966 flags.go:64] FLAG: --kubelet-cgroups="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578545 4966 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578552 4966 flags.go:64] FLAG: --lock-file="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578557 4966 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578562 4966 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578600 4966 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578608 4966 flags.go:64] FLAG: --log-json-split-stream="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578808 4966 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578820 4966 flags.go:64] FLAG: --log-text-split-stream="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578826 4966 flags.go:64] FLAG: --logging-format="text" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578830 4966 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578836 4966 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578841 4966 flags.go:64] FLAG: --manifest-url="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578845 4966 flags.go:64] FLAG: --manifest-url-header="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578863 4966 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578868 4966 flags.go:64] FLAG: --max-open-files="1000000" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578896 4966 flags.go:64] FLAG: --max-pods="110" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578903 4966 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578909 4966 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578914 4966 flags.go:64] FLAG: --memory-manager-policy="None" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578920 4966 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578925 4966 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578931 4966 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578937 4966 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578954 4966 flags.go:64] FLAG: --node-status-max-images="50" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578959 4966 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578964 4966 flags.go:64] FLAG: --oom-score-adj="-999" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578968 4966 flags.go:64] FLAG: --pod-cidr="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578972 4966 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578981 4966 flags.go:64] FLAG: --pod-manifest-path="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578986 4966 flags.go:64] FLAG: --pod-max-pids="-1" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578990 4966 flags.go:64] FLAG: --pods-per-core="0" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.578995 4966 flags.go:64] FLAG: --port="10250" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579000 4966 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579004 4966 flags.go:64] FLAG: --provider-id="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579009 4966 flags.go:64] FLAG: --qos-reserved="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579014 4966 flags.go:64] FLAG: --read-only-port="10255" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579018 4966 flags.go:64] FLAG: --register-node="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579023 4966 flags.go:64] FLAG: --register-schedulable="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579032 4966 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579046 4966 flags.go:64] FLAG: --registry-burst="10" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579052 4966 flags.go:64] FLAG: --registry-qps="5" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579058 4966 flags.go:64] FLAG: --reserved-cpus="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579064 4966 flags.go:64] FLAG: --reserved-memory="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579071 4966 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579077 4966 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579082 4966 flags.go:64] FLAG: --rotate-certificates="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579086 4966 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579090 4966 flags.go:64] FLAG: --runonce="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579094 4966 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579099 4966 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579103 4966 flags.go:64] FLAG: --seccomp-default="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579107 4966 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579111 4966 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579116 4966 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579120 4966 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579125 4966 flags.go:64] FLAG: --storage-driver-password="root" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579129 4966 flags.go:64] FLAG: --storage-driver-secure="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579133 4966 flags.go:64] FLAG: --storage-driver-table="stats" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579137 4966 flags.go:64] FLAG: --storage-driver-user="root" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579141 4966 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579145 4966 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579150 4966 flags.go:64] FLAG: --system-cgroups="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579154 4966 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579162 4966 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579166 4966 flags.go:64] FLAG: --tls-cert-file="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579171 4966 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579177 4966 flags.go:64] FLAG: --tls-min-version="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579181 4966 flags.go:64] FLAG: --tls-private-key-file="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579185 4966 flags.go:64] FLAG: --topology-manager-policy="none" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579189 4966 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579194 4966 flags.go:64] FLAG: --topology-manager-scope="container" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579199 4966 flags.go:64] FLAG: --v="2" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579204 4966 flags.go:64] FLAG: --version="false" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579210 4966 flags.go:64] FLAG: --vmodule="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579216 4966 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579221 4966 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579336 4966 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579340 4966 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579345 4966 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579349 4966 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579353 4966 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579357 4966 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579360 4966 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579364 4966 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579367 4966 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579371 4966 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579375 4966 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579379 4966 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579382 4966 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579385 4966 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579389 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579392 4966 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579396 4966 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579400 4966 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579404 4966 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579408 4966 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579412 4966 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579415 4966 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579419 4966 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579423 4966 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579429 4966 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579433 4966 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579437 4966 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579441 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579446 4966 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579451 4966 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579456 4966 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579460 4966 feature_gate.go:330] unrecognized feature gate: Example Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579464 4966 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579468 4966 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579472 4966 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579476 4966 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579480 4966 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579483 4966 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579493 4966 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579496 4966 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579500 4966 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579503 4966 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579507 4966 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579511 4966 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579514 4966 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579518 4966 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579522 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579526 4966 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579529 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579533 4966 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579536 4966 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579539 4966 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579543 4966 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579546 4966 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579550 4966 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579553 4966 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579557 4966 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579560 4966 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579565 4966 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579569 4966 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579574 4966 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579578 4966 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579581 4966 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579585 4966 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579589 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579592 4966 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579596 4966 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579599 4966 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579603 4966 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579607 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.579610 4966 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.579616 4966 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.586372 4966 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.586408 4966 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586516 4966 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586529 4966 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586534 4966 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586538 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586544 4966 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586548 4966 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586553 4966 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586558 4966 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586563 4966 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586567 4966 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586571 4966 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586576 4966 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586579 4966 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586583 4966 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586587 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586591 4966 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586595 4966 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586599 4966 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586603 4966 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586606 4966 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586611 4966 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586616 4966 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586620 4966 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586624 4966 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586628 4966 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586632 4966 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586637 4966 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586642 4966 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586646 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586650 4966 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586653 4966 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586657 4966 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586660 4966 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586664 4966 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586670 4966 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586673 4966 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586678 4966 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586683 4966 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586687 4966 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586691 4966 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586694 4966 feature_gate.go:330] unrecognized feature gate: Example Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586698 4966 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586702 4966 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586706 4966 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586709 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586712 4966 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586716 4966 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586719 4966 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586723 4966 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586726 4966 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586730 4966 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586733 4966 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586737 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586740 4966 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586744 4966 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586748 4966 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586751 4966 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586755 4966 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586758 4966 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586762 4966 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586766 4966 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586769 4966 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586774 4966 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586778 4966 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586782 4966 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586785 4966 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586789 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586793 4966 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586798 4966 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586802 4966 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586817 4966 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.586824 4966 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586978 4966 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586987 4966 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586992 4966 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.586996 4966 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587000 4966 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587004 4966 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587008 4966 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587011 4966 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587015 4966 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587019 4966 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587023 4966 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587027 4966 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587031 4966 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587035 4966 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587039 4966 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587043 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587047 4966 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587050 4966 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587054 4966 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587057 4966 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587061 4966 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587065 4966 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587068 4966 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587073 4966 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587079 4966 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587084 4966 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587089 4966 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587093 4966 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587097 4966 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587101 4966 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587105 4966 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587109 4966 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587115 4966 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587118 4966 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587123 4966 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587127 4966 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587131 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587135 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587140 4966 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587144 4966 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587148 4966 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587152 4966 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587156 4966 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587160 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587164 4966 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587168 4966 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587172 4966 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587175 4966 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587179 4966 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587182 4966 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587186 4966 feature_gate.go:330] unrecognized feature gate: Example Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587190 4966 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587194 4966 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587198 4966 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587201 4966 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587205 4966 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587209 4966 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587213 4966 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587216 4966 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587220 4966 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587224 4966 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587227 4966 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587232 4966 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587236 4966 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587242 4966 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587246 4966 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587251 4966 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587255 4966 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587261 4966 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587267 4966 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.587274 4966 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.587281 4966 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.587623 4966 server.go:940] "Client rotation is on, will bootstrap in background" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.589961 4966 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.590049 4966 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.590573 4966 server.go:997] "Starting client certificate rotation" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.590588 4966 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.590912 4966 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-12 08:23:17.828266848 +0000 UTC Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.591140 4966 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 624h2m13.237128336s for next certificate rotation Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.595417 4966 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.596897 4966 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.604070 4966 log.go:25] "Validated CRI v1 runtime API" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.616782 4966 log.go:25] "Validated CRI v1 image API" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.619399 4966 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.621744 4966 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-17-08-16-17-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.621781 4966 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.637200 4966 manager.go:217] Machine: {Timestamp:2025-12-17 08:21:04.635342182 +0000 UTC m=+0.180412214 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199476736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:34a07af6-4eb8-4386-b056-0160b9517aa1 BootID:dbf2ad73-4183-43fc-86d4-935121ab90f9 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8a:07:21 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8a:07:21 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:6a:e2:0d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5c:6f:c8 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:bf:77:94 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:64:b7:3b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ca:0a:da:57:13:d8 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a2:a5:9b:a4:f0:07 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199476736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.637557 4966 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.637824 4966 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.638560 4966 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.638764 4966 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.638803 4966 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.638999 4966 topology_manager.go:138] "Creating topology manager with none policy" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.639009 4966 container_manager_linux.go:303] "Creating device plugin manager" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.639216 4966 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.639251 4966 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.639503 4966 state_mem.go:36] "Initialized new in-memory state store" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.639720 4966 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.640381 4966 kubelet.go:418] "Attempting to sync node with API server" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.640399 4966 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.640421 4966 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.640432 4966 kubelet.go:324] "Adding apiserver pod source" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.640442 4966 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.642050 4966 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.642365 4966 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643107 4966 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.643280 4966 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.643270 4966 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.643393 4966 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.643429 4966 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643686 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643705 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643712 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643718 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643729 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643736 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643742 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643752 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643759 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643766 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643782 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.643789 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.644252 4966 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.644687 4966 server.go:1280] "Started kubelet" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.645393 4966 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.645400 4966 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.646022 4966 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.646204 4966 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.646244 4966 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 17 08:21:04 crc systemd[1]: Started Kubernetes Kubelet. Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.646404 4966 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 09:26:11.75378941 +0000 UTC Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.646626 4966 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.646638 4966 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.646730 4966 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.647573 4966 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.648284 4966 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.654743 4966 server.go:460] "Adding debug handlers to kubelet server" Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.654466 4966 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1881f2f562d01814 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-17 08:21:04.64465922 +0000 UTC m=+0.189729161,LastTimestamp:2025-12-17 08:21:04.64465922 +0000 UTC m=+0.189729161,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.656267 4966 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.656315 4966 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.656436 4966 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.656458 4966 factory.go:55] Registering systemd factory Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.656481 4966 factory.go:221] Registration of the systemd container factory successfully Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.654152 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="200ms" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.656722 4966 factory.go:153] Registering CRI-O factory Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.656739 4966 factory.go:221] Registration of the crio container factory successfully Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.656802 4966 factory.go:103] Registering Raw factory Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.656821 4966 manager.go:1196] Started watching for new ooms in manager Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.657422 4966 manager.go:319] Starting recovery of all containers Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665471 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665514 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665531 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665551 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665565 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665575 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665588 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665599 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665613 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665624 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665634 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665646 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665658 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665671 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665683 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665695 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665730 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665743 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665755 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665766 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665779 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665790 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665801 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665816 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665830 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665867 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665901 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665916 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665930 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665943 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665955 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665967 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665985 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.665997 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666030 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666043 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666056 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666067 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666079 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666091 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666105 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666116 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666128 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666141 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666155 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666168 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666180 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666196 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666209 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666223 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666237 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666249 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666266 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666281 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666296 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666310 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666324 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666337 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666350 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666362 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666375 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666388 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666438 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666454 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666468 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666482 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666494 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666506 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666518 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666530 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666542 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666554 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666568 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666581 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666597 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666610 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666624 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666636 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666671 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666687 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666701 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666713 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666726 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666740 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666751 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666763 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666777 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666790 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666803 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666816 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666829 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666842 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666856 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666884 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666899 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666914 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666928 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666940 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666954 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666967 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666981 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.666996 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667009 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667023 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667041 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667053 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667067 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667080 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667094 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667107 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667122 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667137 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667150 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667162 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667175 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667187 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667199 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667213 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667225 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667236 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667248 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667260 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667272 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667284 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667295 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667307 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667318 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667331 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667344 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667356 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667368 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667381 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667393 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667405 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667417 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667467 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667484 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667497 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667510 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667522 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667536 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667550 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667562 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667577 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667589 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667600 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667613 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667627 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667639 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667653 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667665 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667678 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667689 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667702 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667715 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667727 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667740 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667752 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667763 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667776 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667789 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667800 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667812 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667822 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667833 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667845 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667856 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667867 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667917 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667929 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667942 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667955 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667968 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667980 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.667992 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668006 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668019 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668031 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668043 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668056 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668068 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668080 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668092 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668108 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668120 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668133 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668147 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668160 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668174 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668187 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668200 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668215 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668229 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668241 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668253 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668265 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668278 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668291 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668810 4966 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668832 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668846 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668857 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668885 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668897 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668918 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668929 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668939 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668952 4966 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668963 4966 reconstruct.go:97] "Volume reconstruction finished" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.668969 4966 reconciler.go:26] "Reconciler: start to sync state" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.679348 4966 manager.go:324] Recovery completed Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.686355 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.687510 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.687545 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.687554 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.688393 4966 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.688411 4966 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.688431 4966 state_mem.go:36] "Initialized new in-memory state store" Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.756853 4966 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.818280 4966 policy_none.go:49] "None policy: Start" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.819778 4966 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.819952 4966 state_mem.go:35] "Initializing new in-memory state store" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.827409 4966 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.829098 4966 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.829147 4966 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.829180 4966 kubelet.go:2335] "Starting kubelet main sync loop" Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.829253 4966 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 17 08:21:04 crc kubenswrapper[4966]: W1217 08:21:04.830128 4966 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.830200 4966 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.857701 4966 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.858139 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="400ms" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.871924 4966 manager.go:334] "Starting Device Plugin manager" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.871980 4966 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.871994 4966 server.go:79] "Starting device plugin registration server" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.872332 4966 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.872478 4966 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.872594 4966 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.872710 4966 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.872719 4966 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.880314 4966 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.929697 4966 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.930100 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.931348 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.931380 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.931391 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.931514 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.932021 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.932089 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.932728 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.932776 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.932787 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.932995 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.933180 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.933226 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.933694 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.933717 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.933726 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.934188 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.934224 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.934235 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.935017 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.935129 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.935221 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.935360 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.935778 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.935906 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.936641 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.936745 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.936862 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.937173 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.937297 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.937352 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.937558 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.937585 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.937596 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.938590 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.938616 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.938627 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.938596 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.938924 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.939019 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.938823 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.939342 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.940097 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.940217 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.940323 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.972853 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.973425 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.973538 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.973683 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.973780 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.973805 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.973814 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.973837 4966 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.974068 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.974274 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.974393 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.974528 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.974623 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.974702 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.974788 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: E1217 08:21:04.974289 4966 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.975004 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.975084 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.975187 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.975290 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 17 08:21:04 crc kubenswrapper[4966]: I1217 08:21:04.975392 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076071 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076134 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076167 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076198 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076229 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076257 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076283 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076311 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076341 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076368 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076395 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076421 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076448 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076475 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076504 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076701 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076702 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076854 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076981 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076985 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076959 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076991 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.077005 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.077016 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.077054 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.077027 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.076979 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.077053 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.077430 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.077540 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.174996 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.176482 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.176515 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.176526 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.176547 4966 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 17 08:21:05 crc kubenswrapper[4966]: E1217 08:21:05.177046 4966 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Dec 17 08:21:05 crc kubenswrapper[4966]: E1217 08:21:05.259897 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="800ms" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.266222 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: W1217 08:21:05.289053 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-b13d72146a3adee4357ce20bab732c889a5c4aca3f4f605988fcb2eac1991d74 WatchSource:0}: Error finding container b13d72146a3adee4357ce20bab732c889a5c4aca3f4f605988fcb2eac1991d74: Status 404 returned error can't find the container with id b13d72146a3adee4357ce20bab732c889a5c4aca3f4f605988fcb2eac1991d74 Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.292502 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.316528 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: W1217 08:21:05.318258 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6e50dce2ebcc874bfdd513a5d305c298dda01d94b6210181027074bddd13e056 WatchSource:0}: Error finding container 6e50dce2ebcc874bfdd513a5d305c298dda01d94b6210181027074bddd13e056: Status 404 returned error can't find the container with id 6e50dce2ebcc874bfdd513a5d305c298dda01d94b6210181027074bddd13e056 Dec 17 08:21:05 crc kubenswrapper[4966]: W1217 08:21:05.332007 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a290688a69010aa45d2c37538fca555fc68fe903a8d16cdbb305c5996bb5e404 WatchSource:0}: Error finding container a290688a69010aa45d2c37538fca555fc68fe903a8d16cdbb305c5996bb5e404: Status 404 returned error can't find the container with id a290688a69010aa45d2c37538fca555fc68fe903a8d16cdbb305c5996bb5e404 Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.341490 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.351421 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 17 08:21:05 crc kubenswrapper[4966]: W1217 08:21:05.362035 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-20d9d271cd31e3f672378bac567135a5af25e553a275d3cf4ea2e469d407c766 WatchSource:0}: Error finding container 20d9d271cd31e3f672378bac567135a5af25e553a275d3cf4ea2e469d407c766: Status 404 returned error can't find the container with id 20d9d271cd31e3f672378bac567135a5af25e553a275d3cf4ea2e469d407c766 Dec 17 08:21:05 crc kubenswrapper[4966]: W1217 08:21:05.370686 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-50334b5b80f2167269147efdb32fd67b689961ad7f764914ee9a51a7fce3464c WatchSource:0}: Error finding container 50334b5b80f2167269147efdb32fd67b689961ad7f764914ee9a51a7fce3464c: Status 404 returned error can't find the container with id 50334b5b80f2167269147efdb32fd67b689961ad7f764914ee9a51a7fce3464c Dec 17 08:21:05 crc kubenswrapper[4966]: W1217 08:21:05.447055 4966 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:05 crc kubenswrapper[4966]: E1217 08:21:05.447140 4966 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.577236 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.579848 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.579913 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.579923 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.579948 4966 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 17 08:21:05 crc kubenswrapper[4966]: E1217 08:21:05.580421 4966 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Dec 17 08:21:05 crc kubenswrapper[4966]: W1217 08:21:05.627541 4966 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:05 crc kubenswrapper[4966]: E1217 08:21:05.627938 4966 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.647129 4966 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 04:06:17.835101001 +0000 UTC Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.647174 4966 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 43h45m12.187929936s for next certificate rotation Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.648468 4966 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.836976 4966 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ecbe78e8e0a9eea2977bbf6699d864624a393bd237d4b4682103527fed7f5ed6" exitCode=0 Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.837055 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ecbe78e8e0a9eea2977bbf6699d864624a393bd237d4b4682103527fed7f5ed6"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.837168 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"20d9d271cd31e3f672378bac567135a5af25e553a275d3cf4ea2e469d407c766"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.837275 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.838504 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.838548 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.838586 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.839266 4966 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0" exitCode=0 Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.839311 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.839325 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a290688a69010aa45d2c37538fca555fc68fe903a8d16cdbb305c5996bb5e404"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.839376 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.840207 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.840229 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.840240 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.841695 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.841792 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6e50dce2ebcc874bfdd513a5d305c298dda01d94b6210181027074bddd13e056"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.844991 4966 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466" exitCode=0 Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.845057 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.845083 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b13d72146a3adee4357ce20bab732c889a5c4aca3f4f605988fcb2eac1991d74"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.845180 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.845999 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.846029 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.846074 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.847150 4966 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6" exitCode=0 Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.847180 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.847221 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"50334b5b80f2167269147efdb32fd67b689961ad7f764914ee9a51a7fce3464c"} Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.847314 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.848054 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.848097 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.848114 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.849153 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.850236 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.850260 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:05 crc kubenswrapper[4966]: I1217 08:21:05.850284 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:06 crc kubenswrapper[4966]: E1217 08:21:06.065035 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="1.6s" Dec 17 08:21:06 crc kubenswrapper[4966]: W1217 08:21:06.099142 4966 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:06 crc kubenswrapper[4966]: E1217 08:21:06.099210 4966 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:21:06 crc kubenswrapper[4966]: W1217 08:21:06.114184 4966 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:06 crc kubenswrapper[4966]: E1217 08:21:06.114252 4966 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.413912 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.416143 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.416216 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.416228 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.416249 4966 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 17 08:21:06 crc kubenswrapper[4966]: E1217 08:21:06.416651 4966 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.649251 4966 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.854055 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.854104 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.854117 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.854216 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.855030 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.855059 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.855069 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.856795 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.856849 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.856860 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.857095 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.858519 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.858543 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.858551 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.859392 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.859421 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.859436 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.859449 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.860925 4966 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4" exitCode=0 Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.860979 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.861092 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.863192 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.863216 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.863229 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.867723 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3f3d2e6b22e58d52949ece30cf7287e8e2102b3da6b4859fe50ef6e3f1ff147b"} Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.867856 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.868595 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.868628 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:06 crc kubenswrapper[4966]: I1217 08:21:06.868638 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.872739 4966 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b" exitCode=0 Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.872817 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b"} Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.872978 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.874105 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.874150 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.874186 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.877466 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.877486 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7"} Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.877473 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.878283 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.878313 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.878312 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.878324 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.878332 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:07 crc kubenswrapper[4966]: I1217 08:21:07.878340 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.017627 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.018928 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.018992 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.019007 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.019052 4966 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.468148 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.470321 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.887201 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7"} Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.887257 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4"} Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.887276 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.887281 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103"} Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.887438 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.887435 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549"} Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.887591 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6"} Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.888572 4966 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.888637 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.888732 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.888771 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.888788 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.888732 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.888905 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.888941 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.890443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.890533 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:08 crc kubenswrapper[4966]: I1217 08:21:08.890600 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.025923 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.273764 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.282476 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.889589 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.889786 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.889840 4966 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.889937 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.890625 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.890653 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.890665 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.891507 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.891523 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.891538 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.891539 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.891562 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:09 crc kubenswrapper[4966]: I1217 08:21:09.891550 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.059546 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.373789 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.891514 4966 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.892144 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.892212 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.892300 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.893509 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.893598 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.893653 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.893858 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.893918 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.893932 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.894021 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.894068 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:10 crc kubenswrapper[4966]: I1217 08:21:10.894090 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:11 crc kubenswrapper[4966]: I1217 08:21:11.894509 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:11 crc kubenswrapper[4966]: I1217 08:21:11.895926 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:11 crc kubenswrapper[4966]: I1217 08:21:11.895971 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:11 crc kubenswrapper[4966]: I1217 08:21:11.895993 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.101594 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.102109 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.103475 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.103663 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.103829 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.658140 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.658793 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.660458 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.660529 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:12 crc kubenswrapper[4966]: I1217 08:21:12.660540 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:13 crc kubenswrapper[4966]: I1217 08:21:13.059824 4966 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 17 08:21:13 crc kubenswrapper[4966]: I1217 08:21:13.059955 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 08:21:13 crc kubenswrapper[4966]: I1217 08:21:13.106461 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:13 crc kubenswrapper[4966]: I1217 08:21:13.107109 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:13 crc kubenswrapper[4966]: I1217 08:21:13.109375 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:13 crc kubenswrapper[4966]: I1217 08:21:13.109516 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:13 crc kubenswrapper[4966]: I1217 08:21:13.109539 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:14 crc kubenswrapper[4966]: E1217 08:21:14.880484 4966 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 17 08:21:16 crc kubenswrapper[4966]: I1217 08:21:16.667744 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:16 crc kubenswrapper[4966]: I1217 08:21:16.668142 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:16 crc kubenswrapper[4966]: I1217 08:21:16.669121 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:16 crc kubenswrapper[4966]: I1217 08:21:16.669218 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:16 crc kubenswrapper[4966]: I1217 08:21:16.669290 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:16 crc kubenswrapper[4966]: I1217 08:21:16.977294 4966 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 17 08:21:16 crc kubenswrapper[4966]: I1217 08:21:16.977374 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 17 08:21:16 crc kubenswrapper[4966]: I1217 08:21:16.985628 4966 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 17 08:21:16 crc kubenswrapper[4966]: I1217 08:21:16.985922 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 17 08:21:18 crc kubenswrapper[4966]: I1217 08:21:18.475698 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:18 crc kubenswrapper[4966]: I1217 08:21:18.475971 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:18 crc kubenswrapper[4966]: I1217 08:21:18.477407 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:18 crc kubenswrapper[4966]: I1217 08:21:18.477447 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:18 crc kubenswrapper[4966]: I1217 08:21:18.477483 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.380410 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.380650 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.381369 4966 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.381460 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.382017 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.382078 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.382089 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.386592 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.919617 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.921325 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.921530 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.921576 4966 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.922140 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:20 crc kubenswrapper[4966]: I1217 08:21:20.922183 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 17 08:21:21 crc kubenswrapper[4966]: E1217 08:21:21.969951 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 17 08:21:21 crc kubenswrapper[4966]: I1217 08:21:21.974714 4966 trace.go:236] Trace[373856594]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Dec-2025 08:21:08.756) (total time: 13217ms): Dec 17 08:21:21 crc kubenswrapper[4966]: Trace[373856594]: ---"Objects listed" error: 13217ms (08:21:21.974) Dec 17 08:21:21 crc kubenswrapper[4966]: Trace[373856594]: [13.217919855s] [13.217919855s] END Dec 17 08:21:21 crc kubenswrapper[4966]: I1217 08:21:21.974747 4966 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 17 08:21:21 crc kubenswrapper[4966]: I1217 08:21:21.975202 4966 trace.go:236] Trace[344657379]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Dec-2025 08:21:07.723) (total time: 14251ms): Dec 17 08:21:21 crc kubenswrapper[4966]: Trace[344657379]: ---"Objects listed" error: 14251ms (08:21:21.975) Dec 17 08:21:21 crc kubenswrapper[4966]: Trace[344657379]: [14.251243526s] [14.251243526s] END Dec 17 08:21:21 crc kubenswrapper[4966]: I1217 08:21:21.975353 4966 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 17 08:21:21 crc kubenswrapper[4966]: I1217 08:21:21.977212 4966 trace.go:236] Trace[136046764]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Dec-2025 08:21:08.466) (total time: 13510ms): Dec 17 08:21:21 crc kubenswrapper[4966]: Trace[136046764]: ---"Objects listed" error: 13510ms (08:21:21.977) Dec 17 08:21:21 crc kubenswrapper[4966]: Trace[136046764]: [13.510648852s] [13.510648852s] END Dec 17 08:21:21 crc kubenswrapper[4966]: I1217 08:21:21.977231 4966 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 17 08:21:21 crc kubenswrapper[4966]: I1217 08:21:21.978385 4966 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 17 08:21:21 crc kubenswrapper[4966]: I1217 08:21:21.978509 4966 trace.go:236] Trace[1649383949]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Dec-2025 08:21:07.246) (total time: 14732ms): Dec 17 08:21:21 crc kubenswrapper[4966]: Trace[1649383949]: ---"Objects listed" error: 14732ms (08:21:21.978) Dec 17 08:21:21 crc kubenswrapper[4966]: Trace[1649383949]: [14.732238347s] [14.732238347s] END Dec 17 08:21:21 crc kubenswrapper[4966]: I1217 08:21:21.978526 4966 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 17 08:21:21 crc kubenswrapper[4966]: E1217 08:21:21.979650 4966 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.032359 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.041046 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.127650 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.139732 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.651383 4966 apiserver.go:52] "Watching apiserver" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.654791 4966 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.655425 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-etcd/etcd-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.656468 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.656888 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.656949 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.657050 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.657097 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.657174 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.657579 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.657866 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.657941 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.658900 4966 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.658931 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.660981 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.661075 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.661276 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.661274 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.662327 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.662431 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.662378 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.662568 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.662700 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.684704 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.702667 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.717586 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.734968 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.746141 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.747536 4966 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.754496 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.763591 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.772372 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.782472 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.782510 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.782527 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.782546 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.782568 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.782590 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.782608 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.782864 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.782955 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783045 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783073 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783095 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783085 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783109 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783173 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783202 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783232 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783256 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783264 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783279 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783302 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783327 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783347 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783367 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783447 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783469 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783528 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783573 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783600 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783624 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783646 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783669 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783691 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783732 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783754 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783775 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783798 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783820 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783844 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783907 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783947 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783972 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784006 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784052 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784076 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784100 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784122 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784148 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783570 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783602 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783746 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.783819 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784027 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784064 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784092 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.784166 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:21:23.284149727 +0000 UTC m=+18.829219669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784920 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784940 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784962 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784978 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784994 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785014 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785033 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785067 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785084 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785099 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785115 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785132 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785147 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785165 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785182 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785197 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785212 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785227 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785244 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785261 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785276 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785291 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785306 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785322 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785340 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785358 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785375 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785392 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785410 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785426 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785442 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785459 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785475 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785493 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785508 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785523 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785541 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785560 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785587 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785606 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785621 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785637 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785654 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785670 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784990 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785727 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785727 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785008 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784275 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784364 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784346 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784459 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784532 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784599 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784677 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784728 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784741 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784788 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784843 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784855 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785135 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785232 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785270 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.784180 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785307 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785372 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785387 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785392 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785439 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785908 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785539 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785563 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785681 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785705 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785942 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785746 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785980 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.786226 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.786382 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.787286 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.787600 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.787622 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.787668 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.787929 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788000 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788143 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.785747 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788206 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788232 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788255 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788290 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788313 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788334 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788370 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788397 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788424 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788446 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788469 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788531 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788555 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788596 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788620 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788644 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788670 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788694 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788716 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788740 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788763 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788785 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788807 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788830 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788852 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788946 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788960 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.788974 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789031 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789082 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789108 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789130 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789152 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789175 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789198 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789223 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789245 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789267 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789290 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789315 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789338 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789361 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789385 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789410 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789433 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789455 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789479 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789502 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789524 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789548 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789573 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789597 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789620 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789644 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789667 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789691 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789715 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789739 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789764 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789786 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789811 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789836 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789861 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789903 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789925 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789949 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.790313 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.790457 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.789979 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.790575 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791097 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791424 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791507 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791553 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791577 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791603 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791627 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791650 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791672 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791693 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791716 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791738 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791783 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791806 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791828 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791851 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791916 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791946 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791955 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.791970 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792050 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792091 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792178 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792215 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792249 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792289 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792327 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792361 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792370 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792396 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.792844 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793115 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793147 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793158 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793204 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793254 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793221 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793320 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793555 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793587 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793610 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793636 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793659 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793683 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793707 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793730 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793752 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793784 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793807 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793859 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793905 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.794066 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797077 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797147 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797286 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797329 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797365 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797401 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797441 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797481 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797523 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797583 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797650 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797799 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797833 4966 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797855 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797896 4966 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797916 4966 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797959 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797990 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798009 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798130 4966 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798198 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798214 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798227 4966 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798240 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798252 4966 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798265 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798276 4966 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798323 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798433 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798453 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798466 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798479 4966 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798491 4966 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798505 4966 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798530 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798553 4966 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798573 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798591 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798611 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798624 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798637 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798648 4966 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798661 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798673 4966 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798685 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798698 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798718 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798734 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798746 4966 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798759 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798771 4966 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798793 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798806 4966 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798818 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798830 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798842 4966 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798854 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798928 4966 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799059 4966 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799077 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799091 4966 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799123 4966 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799137 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799202 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799232 4966 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799293 4966 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799311 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799326 4966 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799342 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799358 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799399 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799415 4966 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799430 4966 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799446 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799461 4966 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793430 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793663 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.793781 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800560 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800663 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.794938 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800828 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.794948 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.795018 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.795182 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800857 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.795224 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.795310 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.795771 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.796592 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.797455 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.798397 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799036 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799064 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799414 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.799546 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.795681 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800039 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.801014 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800049 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800150 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800330 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800364 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800380 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800391 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.800899 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.801130 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.801129 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.801249 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.801296 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.801526 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.802624 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.802996 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.803626 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.803972 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.804465 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.804596 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.804635 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.804789 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.804855 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.805169 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.805171 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.805356 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.805447 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.805556 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.805585 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.805620 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.806094 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.806327 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.806344 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.807202 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.807400 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.807711 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.807832 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.808037 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.808160 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.808481 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.808643 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.808785 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.809029 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.809179 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.809211 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.809600 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.809918 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.809945 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810218 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810274 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810418 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810553 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810617 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810664 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810755 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810792 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810815 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.810856 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.811040 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.811198 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.811300 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.811524 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.811531 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.811633 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:23.311609229 +0000 UTC m=+18.856679171 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.812032 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.812194 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.812220 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.812432 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.812676 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.812726 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.812984 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.813040 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.813210 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.813433 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.813808 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.813841 4966 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.814037 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.814324 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.814548 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.814590 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.815119 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.815202 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.815754 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.815823 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.816278 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.816593 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.817366 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.817509 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.817843 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.818132 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.818580 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.818996 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.823934 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.824898 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.825265 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.825310 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.825385 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.826251 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.826325 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.826364 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.826720 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.826836 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.826999 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.827173 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.827486 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.827675 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.827942 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.828200 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.828203 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.828259 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.828301 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.828447 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:23.328417298 +0000 UTC m=+18.873487240 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.828939 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.829002 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:23.328986533 +0000 UTC m=+18.874056475 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.829022 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.831020 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.833361 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.833674 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.834096 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.836646 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.839470 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.840241 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.841751 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.842589 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.843923 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.844607 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.845372 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.846513 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.846543 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.846558 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.846618 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:23.346601562 +0000 UTC m=+18.891671504 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.847295 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.848269 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.850233 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.859250 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.859479 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.864589 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.866210 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.867711 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.869302 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.869587 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.870439 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.870961 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.871676 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.873150 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.874023 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.874647 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.876943 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.877588 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.878054 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.878801 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.878831 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.879618 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.879723 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.880066 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.881336 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.881728 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.882801 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.883787 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.884377 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.885326 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.885808 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.886814 4966 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.886995 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.889064 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.890131 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.890656 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.892354 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.893496 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.894052 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.895252 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.896122 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.896689 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.896847 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.897705 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.900097 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.900743 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902629 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902675 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902729 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902740 4966 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902750 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902760 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902768 4966 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902777 4966 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902787 4966 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902795 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902823 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902831 4966 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902838 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902846 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902854 4966 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902862 4966 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902892 4966 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902903 4966 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902918 4966 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902929 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902942 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902942 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902860 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902951 4966 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902998 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903012 4966 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903025 4966 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903036 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903047 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903058 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903069 4966 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903081 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903092 4966 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903103 4966 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903114 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903125 4966 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903136 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903147 4966 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903159 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903170 4966 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903181 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903195 4966 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903208 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903221 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903233 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903244 4966 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903256 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903267 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903278 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903289 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.903495 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.904275 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.904429 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.902767 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905004 4966 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905068 4966 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905193 4966 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905249 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905307 4966 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905279 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905367 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905539 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905556 4966 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905587 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905597 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905607 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.905619 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.906269 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.906594 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.906780 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907125 4966 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907240 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907271 4966 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907284 4966 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907295 4966 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907307 4966 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907320 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907331 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907346 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907358 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907369 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907380 4966 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907391 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907402 4966 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907413 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907424 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907437 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907448 4966 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907459 4966 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907471 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907482 4966 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907494 4966 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907505 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907517 4966 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907528 4966 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907540 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907547 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907553 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907566 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907578 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907590 4966 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907602 4966 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907613 4966 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907629 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907652 4966 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907670 4966 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907685 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907700 4966 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907715 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907730 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907745 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907760 4966 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907775 4966 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907790 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907805 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907820 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907834 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907849 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907867 4966 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907910 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907926 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907940 4966 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907954 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907968 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907982 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.907996 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908011 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908025 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908042 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908057 4966 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908073 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908089 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908104 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908118 4966 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908132 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908146 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908159 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908172 4966 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908187 4966 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908202 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908219 4966 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.908449 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.909022 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.909842 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.912024 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.922641 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.925728 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.927323 4966 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7" exitCode=255 Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.927364 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7"} Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.932662 4966 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.937570 4966 scope.go:117] "RemoveContainer" containerID="101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.939021 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 17 08:21:22 crc kubenswrapper[4966]: E1217 08:21:22.948153 4966 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.955639 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.966607 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.970733 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.977223 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.979536 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 17 08:21:22 crc kubenswrapper[4966]: W1217 08:21:22.982618 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-e805062af0719a11fc438e24c186bedf141c0ad8e48571f7040024994e0d2bf3 WatchSource:0}: Error finding container e805062af0719a11fc438e24c186bedf141c0ad8e48571f7040024994e0d2bf3: Status 404 returned error can't find the container with id e805062af0719a11fc438e24c186bedf141c0ad8e48571f7040024994e0d2bf3 Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.986459 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 17 08:21:22 crc kubenswrapper[4966]: I1217 08:21:22.990682 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.000916 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: W1217 08:21:23.006923 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-091824425cdc58122305629f8b4bf3e3a35b92bb6a43644abbac7b6e679e6e2c WatchSource:0}: Error finding container 091824425cdc58122305629f8b4bf3e3a35b92bb6a43644abbac7b6e679e6e2c: Status 404 returned error can't find the container with id 091824425cdc58122305629f8b4bf3e3a35b92bb6a43644abbac7b6e679e6e2c Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.016754 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: W1217 08:21:23.019308 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-2fa7964f78aa97289425e53575e77390fd7a706a3e84e81f99ac52b2509dc94c WatchSource:0}: Error finding container 2fa7964f78aa97289425e53575e77390fd7a706a3e84e81f99ac52b2509dc94c: Status 404 returned error can't find the container with id 2fa7964f78aa97289425e53575e77390fd7a706a3e84e81f99ac52b2509dc94c Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.034078 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.048532 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.311026 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.311181 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:21:24.311158529 +0000 UTC m=+19.856228471 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.412044 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.412077 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.412097 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.412116 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412198 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412235 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:24.412223764 +0000 UTC m=+19.957293706 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412287 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412306 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:24.412300417 +0000 UTC m=+19.957370359 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412352 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412361 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412370 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412390 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:24.412384239 +0000 UTC m=+19.957454181 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412424 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412433 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412441 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.412458 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:24.412453411 +0000 UTC m=+19.957523353 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.474347 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-dp4dv"] Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.474608 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dp4dv" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.476682 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.476721 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.477007 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.496830 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.516953 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.526108 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.533714 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.544266 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.561379 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.576551 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.587007 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.602940 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.613771 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5ca416e0-2aab-40ba-b42f-a8ed3afbbafb-hosts-file\") pod \"node-resolver-dp4dv\" (UID: \"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\") " pod="openshift-dns/node-resolver-dp4dv" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.613836 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsjb2\" (UniqueName: \"kubernetes.io/projected/5ca416e0-2aab-40ba-b42f-a8ed3afbbafb-kube-api-access-xsjb2\") pod \"node-resolver-dp4dv\" (UID: \"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\") " pod="openshift-dns/node-resolver-dp4dv" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.621827 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.714987 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5ca416e0-2aab-40ba-b42f-a8ed3afbbafb-hosts-file\") pod \"node-resolver-dp4dv\" (UID: \"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\") " pod="openshift-dns/node-resolver-dp4dv" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.715021 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsjb2\" (UniqueName: \"kubernetes.io/projected/5ca416e0-2aab-40ba-b42f-a8ed3afbbafb-kube-api-access-xsjb2\") pod \"node-resolver-dp4dv\" (UID: \"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\") " pod="openshift-dns/node-resolver-dp4dv" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.715155 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5ca416e0-2aab-40ba-b42f-a8ed3afbbafb-hosts-file\") pod \"node-resolver-dp4dv\" (UID: \"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\") " pod="openshift-dns/node-resolver-dp4dv" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.732150 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsjb2\" (UniqueName: \"kubernetes.io/projected/5ca416e0-2aab-40ba-b42f-a8ed3afbbafb-kube-api-access-xsjb2\") pod \"node-resolver-dp4dv\" (UID: \"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\") " pod="openshift-dns/node-resolver-dp4dv" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.786092 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dp4dv" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.829491 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.829808 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.830004 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.830173 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.879560 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-4q4sn"] Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.880150 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-dxggm"] Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.880643 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.880748 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.884339 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6lq7n"] Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.885144 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-lqk4j"] Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.885371 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lqk4j" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.885499 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:23 crc kubenswrapper[4966]: W1217 08:21:23.891962 4966 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": failed to list *v1.Secret: secrets "ovn-node-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.892008 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-node-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.892101 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.892102 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.892370 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.892501 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.893310 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.893466 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.898972 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.899204 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 17 08:21:23 crc kubenswrapper[4966]: W1217 08:21:23.899342 4966 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.899373 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 17 08:21:23 crc kubenswrapper[4966]: W1217 08:21:23.899473 4966 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.899490 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.899708 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.899852 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 17 08:21:23 crc kubenswrapper[4966]: W1217 08:21:23.900389 4966 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-pwtwl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.900404 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-pwtwl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-node-dockercfg-pwtwl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 17 08:21:23 crc kubenswrapper[4966]: W1217 08:21:23.900443 4966 reflector.go:561] object-"openshift-ovn-kubernetes"/"env-overrides": failed to list *v1.ConfigMap: configmaps "env-overrides" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.900456 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"env-overrides\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"env-overrides\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 17 08:21:23 crc kubenswrapper[4966]: W1217 08:21:23.900555 4966 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovnkube-config": failed to list *v1.ConfigMap: configmaps "ovnkube-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.900573 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnkube-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 17 08:21:23 crc kubenswrapper[4966]: W1217 08:21:23.914219 4966 reflector.go:561] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 17 08:21:23 crc kubenswrapper[4966]: E1217 08:21:23.914265 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.914232 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.914306 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.937766 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.939699 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:23Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.940281 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc"} Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.940979 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.958476 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2fa7964f78aa97289425e53575e77390fd7a706a3e84e81f99ac52b2509dc94c"} Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.967614 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e"} Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.967662 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e805062af0719a11fc438e24c186bedf141c0ad8e48571f7040024994e0d2bf3"} Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.979917 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628"} Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.979963 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1"} Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.979979 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"091824425cdc58122305629f8b4bf3e3a35b92bb6a43644abbac7b6e679e6e2c"} Dec 17 08:21:23 crc kubenswrapper[4966]: I1217 08:21:23.981990 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dp4dv" event={"ID":"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb","Type":"ContainerStarted","Data":"3799daf2bc9b7aa8d2fe85fb68cc8c7f6c02c15a4aec85620ee59aa3e691bfbc"} Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017110 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-cnibin\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017150 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-systemd-units\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017189 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-cnibin\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017209 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-hostroot\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017244 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-log-socket\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017341 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-config\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017423 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-var-lib-cni-multus\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017505 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-var-lib-kubelet\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017580 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017624 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9k9d\" (UniqueName: \"kubernetes.io/projected/1ab08e5a-5118-46ba-9b61-d0e5939c3083-kube-api-access-v9k9d\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017649 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-env-overrides\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017668 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-etc-kubernetes\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017713 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-ovn-kubernetes\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017735 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-run-k8s-cni-cncf-io\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017776 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-conf-dir\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017822 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1ab08e5a-5118-46ba-9b61-d0e5939c3083-cni-binary-copy\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017859 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f703caec-d8d0-4e72-b58a-987e69356984-proxy-tls\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017903 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-etc-openvswitch\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017920 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac204c30-92c9-4b1d-9cda-bef16b7a0008-cni-binary-copy\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017941 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f703caec-d8d0-4e72-b58a-987e69356984-rootfs\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017981 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx2dd\" (UniqueName: \"kubernetes.io/projected/f703caec-d8d0-4e72-b58a-987e69356984-kube-api-access-lx2dd\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.017999 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-netns\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018019 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-os-release\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018034 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-systemd\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018064 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-var-lib-openvswitch\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018079 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-node-log\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018095 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-cni-dir\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018108 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-daemon-config\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018140 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-system-cni-dir\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018156 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-openvswitch\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018170 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovn-node-metrics-cert\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018216 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-script-lib\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018231 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-run-multus-certs\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018246 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f703caec-d8d0-4e72-b58a-987e69356984-mcd-auth-proxy-config\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018261 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-socket-dir-parent\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018290 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24lmz\" (UniqueName: \"kubernetes.io/projected/ac204c30-92c9-4b1d-9cda-bef16b7a0008-kube-api-access-24lmz\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018307 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-kubelet\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018321 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-netd\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018336 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-var-lib-cni-bin\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018373 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-slash\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018388 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018405 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-run-netns\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018453 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1ab08e5a-5118-46ba-9b61-d0e5939c3083-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018467 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-ovn\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018481 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-bin\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018495 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-os-release\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018530 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnvtq\" (UniqueName: \"kubernetes.io/projected/865dfc88-97a7-4cdb-9564-f46cfadae5dd-kube-api-access-vnvtq\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.018551 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-system-cni-dir\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.020374 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.056523 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.081671 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.097591 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.111969 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119396 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f703caec-d8d0-4e72-b58a-987e69356984-proxy-tls\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119430 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-etc-openvswitch\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119446 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac204c30-92c9-4b1d-9cda-bef16b7a0008-cni-binary-copy\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119477 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1ab08e5a-5118-46ba-9b61-d0e5939c3083-cni-binary-copy\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119492 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f703caec-d8d0-4e72-b58a-987e69356984-rootfs\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119507 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx2dd\" (UniqueName: \"kubernetes.io/projected/f703caec-d8d0-4e72-b58a-987e69356984-kube-api-access-lx2dd\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119521 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-netns\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119542 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-var-lib-openvswitch\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119555 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-node-log\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119568 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-cni-dir\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119582 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-os-release\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119609 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-systemd\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119635 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-daemon-config\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119649 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-run-multus-certs\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119664 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-system-cni-dir\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119679 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-openvswitch\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119694 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovn-node-metrics-cert\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119709 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-script-lib\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119725 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f703caec-d8d0-4e72-b58a-987e69356984-mcd-auth-proxy-config\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119740 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-socket-dir-parent\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119756 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24lmz\" (UniqueName: \"kubernetes.io/projected/ac204c30-92c9-4b1d-9cda-bef16b7a0008-kube-api-access-24lmz\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119781 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-kubelet\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119799 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-netd\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119824 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-var-lib-cni-bin\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119839 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-slash\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119855 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119891 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-run-netns\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119915 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-os-release\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119937 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1ab08e5a-5118-46ba-9b61-d0e5939c3083-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119953 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-ovn\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119968 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-bin\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.119984 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-system-cni-dir\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120018 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnvtq\" (UniqueName: \"kubernetes.io/projected/865dfc88-97a7-4cdb-9564-f46cfadae5dd-kube-api-access-vnvtq\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120017 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-socket-dir-parent\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120057 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-systemd\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120017 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-system-cni-dir\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120066 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-cnibin\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120092 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-os-release\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120033 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-cnibin\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120238 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-openvswitch\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120336 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-run-multus-certs\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120375 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-systemd-units\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120399 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-cnibin\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120420 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-hostroot\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120649 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-log-socket\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120672 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-config\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120693 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-var-lib-cni-multus\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120715 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-var-lib-kubelet\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120735 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-etc-kubernetes\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120764 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120786 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9k9d\" (UniqueName: \"kubernetes.io/projected/1ab08e5a-5118-46ba-9b61-d0e5939c3083-kube-api-access-v9k9d\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120811 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-env-overrides\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120904 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-ovn-kubernetes\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120929 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-run-k8s-cni-cncf-io\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120950 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-conf-dir\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.120990 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-system-cni-dir\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121017 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-netd\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121026 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-conf-dir\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121051 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f703caec-d8d0-4e72-b58a-987e69356984-rootfs\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121072 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-bin\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121238 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-var-lib-cni-bin\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121260 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-node-log\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121372 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-cni-dir\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121405 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-var-lib-openvswitch\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121429 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-netns\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121449 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121460 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-ovn\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121629 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-slash\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121644 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-os-release\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121841 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-run-netns\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121839 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac204c30-92c9-4b1d-9cda-bef16b7a0008-cni-binary-copy\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121893 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-kubelet\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121894 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-systemd-units\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121925 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-cnibin\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121954 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-var-lib-kubelet\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.121974 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-etc-openvswitch\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122183 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1ab08e5a-5118-46ba-9b61-d0e5939c3083-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122212 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-etc-kubernetes\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122231 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-hostroot\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122248 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-log-socket\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122292 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-var-lib-cni-multus\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122306 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1ab08e5a-5118-46ba-9b61-d0e5939c3083-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122336 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-script-lib\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122346 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f703caec-d8d0-4e72-b58a-987e69356984-mcd-auth-proxy-config\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122598 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-ovn-kubernetes\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.122643 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ac204c30-92c9-4b1d-9cda-bef16b7a0008-host-run-k8s-cni-cncf-io\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.123924 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1ab08e5a-5118-46ba-9b61-d0e5939c3083-cni-binary-copy\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.125788 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f703caec-d8d0-4e72-b58a-987e69356984-proxy-tls\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.132895 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.139479 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx2dd\" (UniqueName: \"kubernetes.io/projected/f703caec-d8d0-4e72-b58a-987e69356984-kube-api-access-lx2dd\") pod \"machine-config-daemon-dxggm\" (UID: \"f703caec-d8d0-4e72-b58a-987e69356984\") " pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.141902 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9k9d\" (UniqueName: \"kubernetes.io/projected/1ab08e5a-5118-46ba-9b61-d0e5939c3083-kube-api-access-v9k9d\") pod \"multus-additional-cni-plugins-4q4sn\" (UID: \"1ab08e5a-5118-46ba-9b61-d0e5939c3083\") " pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.144397 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24lmz\" (UniqueName: \"kubernetes.io/projected/ac204c30-92c9-4b1d-9cda-bef16b7a0008-kube-api-access-24lmz\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.146266 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.161645 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.169504 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.178765 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.188987 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.199534 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.205106 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.207948 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" Dec 17 08:21:24 crc kubenswrapper[4966]: W1217 08:21:24.215007 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf703caec_d8d0_4e72_b58a_987e69356984.slice/crio-b876c1d915e8a92a61fd8c76ef3cfead9cdad05acbfdba324a321a2814675d95 WatchSource:0}: Error finding container b876c1d915e8a92a61fd8c76ef3cfead9cdad05acbfdba324a321a2814675d95: Status 404 returned error can't find the container with id b876c1d915e8a92a61fd8c76ef3cfead9cdad05acbfdba324a321a2814675d95 Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.217527 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.233034 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.244944 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.257278 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.269787 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.285396 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.296580 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.318212 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.322711 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.322902 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:21:26.322856116 +0000 UTC m=+21.867926068 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.332665 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.347796 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.366003 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.390016 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.424146 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.424197 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.424226 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.424280 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.424675 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.424695 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.424705 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.424763 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:26.424749973 +0000 UTC m=+21.969819915 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.425381 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.425528 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:26.425514893 +0000 UTC m=+21.970584835 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.425649 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.425733 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.425786 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.425891 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:26.425863192 +0000 UTC m=+21.970933134 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.426062 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.426150 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:26.42614278 +0000 UTC m=+21.971212722 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.828629 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.829704 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:24 crc kubenswrapper[4966]: E1217 08:21:24.829811 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.833383 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.834666 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-env-overrides\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.845000 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovn-node-metrics-cert\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.851544 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.869392 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.881177 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.897211 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.915489 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.927377 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.942081 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.957749 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.970113 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.984132 4966 generic.go:334] "Generic (PLEG): container finished" podID="1ab08e5a-5118-46ba-9b61-d0e5939c3083" containerID="eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699" exitCode=0 Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.984190 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" event={"ID":"1ab08e5a-5118-46ba-9b61-d0e5939c3083","Type":"ContainerDied","Data":"eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699"} Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.984221 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" event={"ID":"1ab08e5a-5118-46ba-9b61-d0e5939c3083","Type":"ContainerStarted","Data":"e6bfec7d75a0d6fb214b86adc2e1065e189fec579295a5c67dd45cfb59e7f100"} Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.986507 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.994238 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62"} Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.994308 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532"} Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.994320 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"b876c1d915e8a92a61fd8c76ef3cfead9cdad05acbfdba324a321a2814675d95"} Dec 17 08:21:24 crc kubenswrapper[4966]: I1217 08:21:24.994330 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dp4dv" event={"ID":"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb","Type":"ContainerStarted","Data":"8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.006299 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.019477 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.038718 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.083527 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.121369 4966 configmap.go:193] Couldn't get configMap openshift-multus/multus-daemon-config: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.121456 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-daemon-config podName:ac204c30-92c9-4b1d-9cda-bef16b7a0008 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:25.62144018 +0000 UTC m=+21.166510122 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "multus-daemon-config" (UniqueName: "kubernetes.io/configmap/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-daemon-config") pod "multus-lqk4j" (UID: "ac204c30-92c9-4b1d-9cda-bef16b7a0008") : failed to sync configmap cache: timed out waiting for the condition Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.122474 4966 configmap.go:193] Couldn't get configMap openshift-ovn-kubernetes/ovnkube-config: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.122575 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-config podName:865dfc88-97a7-4cdb-9564-f46cfadae5dd nodeName:}" failed. No retries permitted until 2025-12-17 08:21:25.62254956 +0000 UTC m=+21.167619612 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovnkube-config" (UniqueName: "kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-config") pod "ovnkube-node-6lq7n" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd") : failed to sync configmap cache: timed out waiting for the condition Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.125402 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.133614 4966 projected.go:288] Couldn't get configMap openshift-ovn-kubernetes/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.133650 4966 projected.go:194] Error preparing data for projected volume kube-api-access-vnvtq for pod openshift-ovn-kubernetes/ovnkube-node-6lq7n: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.133698 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/865dfc88-97a7-4cdb-9564-f46cfadae5dd-kube-api-access-vnvtq podName:865dfc88-97a7-4cdb-9564-f46cfadae5dd nodeName:}" failed. No retries permitted until 2025-12-17 08:21:25.633683847 +0000 UTC m=+21.178753789 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vnvtq" (UniqueName: "kubernetes.io/projected/865dfc88-97a7-4cdb-9564-f46cfadae5dd-kube-api-access-vnvtq") pod "ovnkube-node-6lq7n" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd") : failed to sync configmap cache: timed out waiting for the condition Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.149354 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.180003 4966 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.181000 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.181738 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.181777 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.181790 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.182414 4966 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.201181 4966 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.201461 4966 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.202745 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.202781 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.202790 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.202806 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.202816 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.212022 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.220144 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.227594 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.231401 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.231434 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.231442 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.231454 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.231463 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.236140 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.247432 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.252140 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.252180 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.252196 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.252211 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.252220 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.270053 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.278917 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.282323 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.283028 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.283064 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.283073 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.283089 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.283099 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.296111 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.298330 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.301437 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.301474 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.301486 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.301503 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.301514 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.309050 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.314206 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.316727 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.316832 4966 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.319358 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.319418 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.319428 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.319443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.319472 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.332969 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.341939 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.349186 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.361488 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.370793 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.371634 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.385317 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.421849 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.421892 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.421901 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.421915 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.421924 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.488180 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.525049 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.525093 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.525109 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.525133 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.525149 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.627228 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.627259 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.627267 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.627283 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.627291 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.636685 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-config\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.636751 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-daemon-config\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.636809 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnvtq\" (UniqueName: \"kubernetes.io/projected/865dfc88-97a7-4cdb-9564-f46cfadae5dd-kube-api-access-vnvtq\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.637312 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-config\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.637553 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac204c30-92c9-4b1d-9cda-bef16b7a0008-multus-daemon-config\") pod \"multus-lqk4j\" (UID: \"ac204c30-92c9-4b1d-9cda-bef16b7a0008\") " pod="openshift-multus/multus-lqk4j" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.642152 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnvtq\" (UniqueName: \"kubernetes.io/projected/865dfc88-97a7-4cdb-9564-f46cfadae5dd-kube-api-access-vnvtq\") pod \"ovnkube-node-6lq7n\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.722036 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lqk4j" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.729479 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.729513 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.729523 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.729539 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.729551 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.734694 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:25 crc kubenswrapper[4966]: W1217 08:21:25.752854 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod865dfc88_97a7_4cdb_9564_f46cfadae5dd.slice/crio-6ea787bbe772e55ac5e4a7d96f33a23f2e2ed2ddc1f3991cc392b93dd87008c0 WatchSource:0}: Error finding container 6ea787bbe772e55ac5e4a7d96f33a23f2e2ed2ddc1f3991cc392b93dd87008c0: Status 404 returned error can't find the container with id 6ea787bbe772e55ac5e4a7d96f33a23f2e2ed2ddc1f3991cc392b93dd87008c0 Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.829758 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.829914 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.829983 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:25 crc kubenswrapper[4966]: E1217 08:21:25.830042 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.831954 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.831979 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.831987 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.832001 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.832009 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.898595 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-b447x"] Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.898987 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.901592 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.901733 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.902489 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.902977 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.917047 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.927696 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.934836 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.935079 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.935197 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.935301 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.935496 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:25Z","lastTransitionTime":"2025-12-17T08:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.938884 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.949718 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.966834 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.979452 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.996471 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483" exitCode=0 Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.996960 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.997831 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"6ea787bbe772e55ac5e4a7d96f33a23f2e2ed2ddc1f3991cc392b93dd87008c0"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.999195 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7"} Dec 17 08:21:25 crc kubenswrapper[4966]: I1217 08:21:25.999700 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.003449 4966 generic.go:334] "Generic (PLEG): container finished" podID="1ab08e5a-5118-46ba-9b61-d0e5939c3083" containerID="fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a" exitCode=0 Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.003528 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" event={"ID":"1ab08e5a-5118-46ba-9b61-d0e5939c3083","Type":"ContainerDied","Data":"fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.005482 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqk4j" event={"ID":"ac204c30-92c9-4b1d-9cda-bef16b7a0008","Type":"ContainerStarted","Data":"6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.005658 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqk4j" event={"ID":"ac204c30-92c9-4b1d-9cda-bef16b7a0008","Type":"ContainerStarted","Data":"0b79cb316b5e5cce82d511a75ea936be608fdd90a4ecd25036a07ed5753d6505"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.012187 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.024797 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.037273 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.037299 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.037307 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.037320 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.037329 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.041056 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vttjv\" (UniqueName: \"kubernetes.io/projected/42cd7025-811b-4b83-b1df-d43919ce03de-kube-api-access-vttjv\") pod \"node-ca-b447x\" (UID: \"42cd7025-811b-4b83-b1df-d43919ce03de\") " pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.041089 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42cd7025-811b-4b83-b1df-d43919ce03de-host\") pod \"node-ca-b447x\" (UID: \"42cd7025-811b-4b83-b1df-d43919ce03de\") " pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.041120 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/42cd7025-811b-4b83-b1df-d43919ce03de-serviceca\") pod \"node-ca-b447x\" (UID: \"42cd7025-811b-4b83-b1df-d43919ce03de\") " pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.043840 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.063029 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.081351 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.096111 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.142123 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/42cd7025-811b-4b83-b1df-d43919ce03de-serviceca\") pod \"node-ca-b447x\" (UID: \"42cd7025-811b-4b83-b1df-d43919ce03de\") " pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.142272 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vttjv\" (UniqueName: \"kubernetes.io/projected/42cd7025-811b-4b83-b1df-d43919ce03de-kube-api-access-vttjv\") pod \"node-ca-b447x\" (UID: \"42cd7025-811b-4b83-b1df-d43919ce03de\") " pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.142357 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42cd7025-811b-4b83-b1df-d43919ce03de-host\") pod \"node-ca-b447x\" (UID: \"42cd7025-811b-4b83-b1df-d43919ce03de\") " pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.142417 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42cd7025-811b-4b83-b1df-d43919ce03de-host\") pod \"node-ca-b447x\" (UID: \"42cd7025-811b-4b83-b1df-d43919ce03de\") " pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.143953 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/42cd7025-811b-4b83-b1df-d43919ce03de-serviceca\") pod \"node-ca-b447x\" (UID: \"42cd7025-811b-4b83-b1df-d43919ce03de\") " pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.146109 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.147884 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.147922 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.147934 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.147954 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.147967 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.158776 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.158855 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vttjv\" (UniqueName: \"kubernetes.io/projected/42cd7025-811b-4b83-b1df-d43919ce03de-kube-api-access-vttjv\") pod \"node-ca-b447x\" (UID: \"42cd7025-811b-4b83-b1df-d43919ce03de\") " pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.173612 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.184017 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.200119 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.217690 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-b447x" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.221523 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: W1217 08:21:26.229532 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42cd7025_811b_4b83_b1df_d43919ce03de.slice/crio-e773761aad16fc4a5738455bc71d98a8d94b1d569d212759e2c6d5159b9f6ad5 WatchSource:0}: Error finding container e773761aad16fc4a5738455bc71d98a8d94b1d569d212759e2c6d5159b9f6ad5: Status 404 returned error can't find the container with id e773761aad16fc4a5738455bc71d98a8d94b1d569d212759e2c6d5159b9f6ad5 Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.248772 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.251797 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.251824 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.251833 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.251845 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.251854 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.263627 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.285507 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.299541 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.315568 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.331739 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.347599 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.347976 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:21:30.347958335 +0000 UTC m=+25.893028277 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.353856 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.354422 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.354434 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.354453 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.354464 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.365373 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.401300 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.444294 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.448847 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.448913 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.448957 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449007 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449096 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449132 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449139 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.449033 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449148 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449107 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:30.449086801 +0000 UTC m=+25.994156743 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449280 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:30.449262707 +0000 UTC m=+25.994332649 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449100 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449294 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:30.449288377 +0000 UTC m=+25.994358319 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449309 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449321 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.449371 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:30.449356249 +0000 UTC m=+25.994426191 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.456398 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.456436 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.456446 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.456463 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.456473 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.486946 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.522163 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:26Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.558991 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.559047 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.559060 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.559078 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.559089 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.662117 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.662394 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.662536 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.662662 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.662774 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.764895 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.765100 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.765190 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.765262 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.765356 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.831070 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:26 crc kubenswrapper[4966]: E1217 08:21:26.831178 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.867924 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.867957 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.867966 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.867982 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.867992 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.970027 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.970057 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.970066 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.970081 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:26 crc kubenswrapper[4966]: I1217 08:21:26.970089 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:26Z","lastTransitionTime":"2025-12-17T08:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.016915 4966 generic.go:334] "Generic (PLEG): container finished" podID="1ab08e5a-5118-46ba-9b61-d0e5939c3083" containerID="9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb" exitCode=0 Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.017008 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" event={"ID":"1ab08e5a-5118-46ba-9b61-d0e5939c3083","Type":"ContainerDied","Data":"9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.033421 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.033517 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.033541 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.033600 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.033617 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.033634 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.036695 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-b447x" event={"ID":"42cd7025-811b-4b83-b1df-d43919ce03de","Type":"ContainerStarted","Data":"9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.036757 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-b447x" event={"ID":"42cd7025-811b-4b83-b1df-d43919ce03de","Type":"ContainerStarted","Data":"e773761aad16fc4a5738455bc71d98a8d94b1d569d212759e2c6d5159b9f6ad5"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.043270 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.065022 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.072389 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.072427 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.072443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.072465 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.072482 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.084177 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.101768 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.122998 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.140217 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.151220 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.160013 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.173142 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.174531 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.174567 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.174576 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.174590 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.174602 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.183699 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.209079 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.218945 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.229208 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.241794 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.267225 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.276787 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.276830 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.276842 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.276859 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.276883 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.279763 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.293042 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.306614 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.319917 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.333403 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.364516 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.379493 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.379524 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.379532 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.379545 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.379554 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.399796 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.441028 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.481194 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.481231 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.481241 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.481255 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.481265 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.482744 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.521001 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.587488 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.587528 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.587537 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.587553 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.587563 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.592922 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.613006 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.644997 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.686055 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.689669 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.689831 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.689940 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.690032 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.690096 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.722098 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.793102 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.793136 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.793149 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.793165 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.793178 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.830378 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:27 crc kubenswrapper[4966]: E1217 08:21:27.830503 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.830830 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:27 crc kubenswrapper[4966]: E1217 08:21:27.830910 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.895660 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.895849 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.895863 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.895907 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.895920 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.997637 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.997922 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.998331 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.998361 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:27 crc kubenswrapper[4966]: I1217 08:21:27.998381 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:27Z","lastTransitionTime":"2025-12-17T08:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.040898 4966 generic.go:334] "Generic (PLEG): container finished" podID="1ab08e5a-5118-46ba-9b61-d0e5939c3083" containerID="effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc" exitCode=0 Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.040948 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" event={"ID":"1ab08e5a-5118-46ba-9b61-d0e5939c3083","Type":"ContainerDied","Data":"effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.058003 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.072927 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.083944 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.098006 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.100212 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.100236 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.100250 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.100263 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.100271 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:28Z","lastTransitionTime":"2025-12-17T08:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.108945 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.119765 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.128362 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.139707 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.155113 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.169623 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.179306 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.201378 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.201928 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.201959 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.201969 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.201985 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.201998 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:28Z","lastTransitionTime":"2025-12-17T08:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.241672 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.285750 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.303768 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.303808 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.303820 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.303835 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.303847 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:28Z","lastTransitionTime":"2025-12-17T08:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.327487 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.406210 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.406256 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.406267 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.406282 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.406293 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:28Z","lastTransitionTime":"2025-12-17T08:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.508800 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.508836 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.508848 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.508890 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.508905 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:28Z","lastTransitionTime":"2025-12-17T08:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.611439 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.611479 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.611487 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.611502 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.611513 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:28Z","lastTransitionTime":"2025-12-17T08:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.713610 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.713641 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.713652 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.713669 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.713679 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:28Z","lastTransitionTime":"2025-12-17T08:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.815803 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.815836 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.815845 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.815861 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.815886 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:28Z","lastTransitionTime":"2025-12-17T08:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.864204 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:28 crc kubenswrapper[4966]: E1217 08:21:28.864374 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.918367 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.918400 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.918409 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.918424 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:28 crc kubenswrapper[4966]: I1217 08:21:28.918432 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:28Z","lastTransitionTime":"2025-12-17T08:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.021617 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.021691 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.021707 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.021741 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.021758 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.047465 4966 generic.go:334] "Generic (PLEG): container finished" podID="1ab08e5a-5118-46ba-9b61-d0e5939c3083" containerID="58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e" exitCode=0 Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.047532 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" event={"ID":"1ab08e5a-5118-46ba-9b61-d0e5939c3083","Type":"ContainerDied","Data":"58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.052764 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.067749 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.078022 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.094203 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.106478 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.114735 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.125650 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.125702 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.125715 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.125735 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.125748 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.137501 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.157153 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.170126 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.185498 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.204430 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.225538 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.229275 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.229365 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.229381 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.229403 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.229420 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.239931 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.254679 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.274039 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.296676 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:29Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.331579 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.331610 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.331618 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.331633 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.331642 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.434083 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.434112 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.434120 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.434133 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.434142 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.536520 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.536555 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.536567 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.536585 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.536598 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.638891 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.638918 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.638928 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.638941 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.638949 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.741208 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.741239 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.741248 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.741264 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.741274 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.831107 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:29 crc kubenswrapper[4966]: E1217 08:21:29.831218 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.831514 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:29 crc kubenswrapper[4966]: E1217 08:21:29.831570 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.843248 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.843276 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.843284 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.843299 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.843307 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.945808 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.945846 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.945860 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.945902 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:29 crc kubenswrapper[4966]: I1217 08:21:29.945914 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:29Z","lastTransitionTime":"2025-12-17T08:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.048206 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.048256 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.048274 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.048302 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.048323 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.059598 4966 generic.go:334] "Generic (PLEG): container finished" podID="1ab08e5a-5118-46ba-9b61-d0e5939c3083" containerID="fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09" exitCode=0 Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.059655 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" event={"ID":"1ab08e5a-5118-46ba-9b61-d0e5939c3083","Type":"ContainerDied","Data":"fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.073443 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.092480 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.104213 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.117560 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.129326 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.143447 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.150259 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.150342 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.150350 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.150363 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.150372 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.165461 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.176341 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.187297 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.203304 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.227745 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.240135 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.259749 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.260084 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.259944 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.260094 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.260254 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.260265 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.277696 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.290527 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.362549 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.362587 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.362599 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.362618 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.362631 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.384470 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.384632 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:21:38.38460695 +0000 UTC m=+33.929676912 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.465082 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.465106 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.465113 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.465128 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.465136 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.486164 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.486218 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.486255 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.486290 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486366 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486405 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486417 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486453 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486464 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:38.486440746 +0000 UTC m=+34.031510768 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486471 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486506 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:38.486488637 +0000 UTC m=+34.031558669 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486425 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486530 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486508 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486590 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:38.486566589 +0000 UTC m=+34.031636541 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.486608 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:38.48660005 +0000 UTC m=+34.031670122 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.568255 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.568295 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.568308 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.568326 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.568340 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.670603 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.670643 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.670657 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.670678 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.670694 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.777554 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.777628 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.777647 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.777672 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.777690 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.830031 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:30 crc kubenswrapper[4966]: E1217 08:21:30.830224 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.879792 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.879855 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.879896 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.879928 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.879948 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.982562 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.982614 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.982630 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.982655 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:30 crc kubenswrapper[4966]: I1217 08:21:30.982672 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:30Z","lastTransitionTime":"2025-12-17T08:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.084974 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.085036 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.085060 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.085088 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.085109 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:31Z","lastTransitionTime":"2025-12-17T08:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.188759 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.188798 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.188808 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.188823 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.188834 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:31Z","lastTransitionTime":"2025-12-17T08:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.291301 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.291332 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.291343 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.291360 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.291372 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:31Z","lastTransitionTime":"2025-12-17T08:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.393918 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.393942 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.393949 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.393964 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.393974 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:31Z","lastTransitionTime":"2025-12-17T08:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.496777 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.496823 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.496832 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.496845 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.496855 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:31Z","lastTransitionTime":"2025-12-17T08:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.598812 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.598855 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.598880 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.598899 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.598910 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:31Z","lastTransitionTime":"2025-12-17T08:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.701563 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.701599 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.701610 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.701630 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.701641 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:31Z","lastTransitionTime":"2025-12-17T08:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.804048 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.804078 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.804089 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.804104 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.804113 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:31Z","lastTransitionTime":"2025-12-17T08:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.829816 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:31 crc kubenswrapper[4966]: E1217 08:21:31.830034 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.830148 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:31 crc kubenswrapper[4966]: E1217 08:21:31.830269 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.905977 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.906029 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.906042 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.906085 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:31 crc kubenswrapper[4966]: I1217 08:21:31.906098 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:31Z","lastTransitionTime":"2025-12-17T08:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.007562 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.007597 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.007606 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.007619 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.007629 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.069172 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" event={"ID":"1ab08e5a-5118-46ba-9b61-d0e5939c3083","Type":"ContainerStarted","Data":"52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.073692 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.074399 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.074471 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.082289 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.095388 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.096598 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.098987 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.107234 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.109584 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.109605 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.109614 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.109628 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.109636 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.119222 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.128537 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.141370 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.163413 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.177154 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.193864 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.205422 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.211734 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.211763 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.211773 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.211790 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.211800 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.223596 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.240561 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.255578 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.277571 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.295393 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.309844 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.313578 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.313613 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.313623 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.313639 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.313649 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.324681 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.335583 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.348380 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.363451 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.376528 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.386800 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.398526 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.409729 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.415738 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.415763 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.415773 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.415787 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.415795 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.421052 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.439120 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.449307 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.468567 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.481814 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.497963 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:32Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.518624 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.518696 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.518708 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.518726 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.518942 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.620968 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.621011 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.621019 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.621034 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.621043 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.723475 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.723523 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.723533 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.723549 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.723561 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.825493 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.825534 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.825545 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.825563 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.825574 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.829910 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:32 crc kubenswrapper[4966]: E1217 08:21:32.830054 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.928285 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.928366 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.928379 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.928396 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:32 crc kubenswrapper[4966]: I1217 08:21:32.928408 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:32Z","lastTransitionTime":"2025-12-17T08:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.031781 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.031830 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.031841 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.031862 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.031889 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.075899 4966 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.134179 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.134206 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.134214 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.134226 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.134235 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.237291 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.238228 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.238265 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.238282 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.238301 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.340647 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.340898 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.340910 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.340928 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.340940 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.444383 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.444428 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.444441 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.444456 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.444467 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.546720 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.546756 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.546767 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.546783 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.546793 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.649587 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.649645 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.649665 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.649689 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.649707 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.753410 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.753465 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.753476 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.753494 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.753505 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.830054 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:33 crc kubenswrapper[4966]: E1217 08:21:33.830207 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.830065 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:33 crc kubenswrapper[4966]: E1217 08:21:33.830351 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.856093 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.856148 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.856163 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.856185 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.856202 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.958074 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.958113 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.958122 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.958138 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:33 crc kubenswrapper[4966]: I1217 08:21:33.958147 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:33Z","lastTransitionTime":"2025-12-17T08:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.061252 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.061297 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.061310 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.061330 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.061340 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.081194 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/0.log" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.085117 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47" exitCode=1 Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.085174 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.086089 4966 scope.go:117] "RemoveContainer" containerID="7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.101232 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.122909 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.135382 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.148375 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.158964 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.166579 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.166607 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.166618 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.166634 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.166648 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.171034 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.183925 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.194998 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.216531 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.232524 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.246041 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.259458 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.269354 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.269389 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.269401 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.269421 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.269437 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.277948 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:33Z\\\",\\\"message\\\":\\\"event handler 7 for removal\\\\nI1217 08:21:33.352067 6165 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1217 08:21:33.352079 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1217 08:21:33.352089 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1217 08:21:33.352102 6165 factory.go:656] Stopping watch factory\\\\nI1217 08:21:33.352116 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1217 08:21:33.352128 6165 handler.go:208] Removed *v1.Node event handler 2\\\\nI1217 08:21:33.352135 6165 handler.go:208] Removed *v1.Node event handler 7\\\\nI1217 08:21:33.352142 6165 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1217 08:21:33.352149 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1217 08:21:33.352268 6165 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352371 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352445 6165 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352725 6165 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.291926 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.306023 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.371001 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.371023 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.371033 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.371050 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.371061 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.473128 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.473165 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.473175 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.473191 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.473202 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.575508 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.575557 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.575571 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.575595 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.575608 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.677431 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.677469 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.677478 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.677492 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.677503 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.779707 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.779747 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.779760 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.779776 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.779785 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.830483 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:34 crc kubenswrapper[4966]: E1217 08:21:34.830640 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.849953 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.864791 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.877177 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.881124 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.881174 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.881186 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.881203 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.881213 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.896185 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.912554 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.922579 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.936212 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.947391 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.967210 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.983614 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.983662 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.983678 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.983700 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.983717 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:34Z","lastTransitionTime":"2025-12-17T08:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.984063 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:34 crc kubenswrapper[4966]: I1217 08:21:34.995330 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:34Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.008171 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.027651 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:33Z\\\",\\\"message\\\":\\\"event handler 7 for removal\\\\nI1217 08:21:33.352067 6165 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1217 08:21:33.352079 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1217 08:21:33.352089 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1217 08:21:33.352102 6165 factory.go:656] Stopping watch factory\\\\nI1217 08:21:33.352116 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1217 08:21:33.352128 6165 handler.go:208] Removed *v1.Node event handler 2\\\\nI1217 08:21:33.352135 6165 handler.go:208] Removed *v1.Node event handler 7\\\\nI1217 08:21:33.352142 6165 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1217 08:21:33.352149 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1217 08:21:33.352268 6165 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352371 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352445 6165 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352725 6165 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.038644 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.049249 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.085733 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.085764 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.085772 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.085785 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.085795 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.088787 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/0.log" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.091141 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c"} Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.091335 4966 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.103976 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.116214 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.127247 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.138171 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.146349 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.154601 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.170530 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.179125 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.188524 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.188558 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.188571 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.188593 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.188607 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.194736 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.205677 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.214621 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.223937 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.241341 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:33Z\\\",\\\"message\\\":\\\"event handler 7 for removal\\\\nI1217 08:21:33.352067 6165 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1217 08:21:33.352079 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1217 08:21:33.352089 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1217 08:21:33.352102 6165 factory.go:656] Stopping watch factory\\\\nI1217 08:21:33.352116 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1217 08:21:33.352128 6165 handler.go:208] Removed *v1.Node event handler 2\\\\nI1217 08:21:33.352135 6165 handler.go:208] Removed *v1.Node event handler 7\\\\nI1217 08:21:33.352142 6165 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1217 08:21:33.352149 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1217 08:21:33.352268 6165 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352371 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352445 6165 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352725 6165 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.284426 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.291201 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.291235 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.291246 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.291262 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.291274 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.320929 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.393828 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.393854 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.393863 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.393905 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.393917 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.497412 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.497478 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.497497 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.497522 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.497541 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.581810 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.581858 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.581913 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.581934 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.581947 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: E1217 08:21:35.603254 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.608374 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.608410 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.608421 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.608437 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.608449 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: E1217 08:21:35.630507 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.635695 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.635739 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.635751 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.635769 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.635780 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: E1217 08:21:35.654691 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.658753 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.658791 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.658802 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.658818 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.658829 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: E1217 08:21:35.677641 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.681503 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.681536 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.681547 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.681564 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.681577 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: E1217 08:21:35.695933 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:35Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:35 crc kubenswrapper[4966]: E1217 08:21:35.696106 4966 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.697624 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.697672 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.697683 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.697700 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.697712 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.800383 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.800416 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.800424 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.800438 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.800447 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.829964 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.830025 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:35 crc kubenswrapper[4966]: E1217 08:21:35.830065 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:35 crc kubenswrapper[4966]: E1217 08:21:35.830176 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.902806 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.902848 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.902863 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.902892 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:35 crc kubenswrapper[4966]: I1217 08:21:35.902901 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:35Z","lastTransitionTime":"2025-12-17T08:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.007447 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.007480 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.007490 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.007505 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.007517 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.018358 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk"] Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.020066 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.022691 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.024530 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.093708 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.099571 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/1.log" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.100036 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/0.log" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.102774 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c" exitCode=1 Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.102820 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.102922 4966 scope.go:117] "RemoveContainer" containerID="7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.103583 4966 scope.go:117] "RemoveContainer" containerID="f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c" Dec 17 08:21:36 crc kubenswrapper[4966]: E1217 08:21:36.103763 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.108931 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.113233 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.113266 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.113278 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.113294 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.113306 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.121734 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.133402 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.143899 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/216f8233-5cc0-4733-b27e-69f9312d9a94-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.143939 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/216f8233-5cc0-4733-b27e-69f9312d9a94-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.143965 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/216f8233-5cc0-4733-b27e-69f9312d9a94-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.143985 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n4cv\" (UniqueName: \"kubernetes.io/projected/216f8233-5cc0-4733-b27e-69f9312d9a94-kube-api-access-2n4cv\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.144216 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.152205 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.161669 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.173391 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.181584 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.202221 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:33Z\\\",\\\"message\\\":\\\"event handler 7 for removal\\\\nI1217 08:21:33.352067 6165 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1217 08:21:33.352079 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1217 08:21:33.352089 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1217 08:21:33.352102 6165 factory.go:656] Stopping watch factory\\\\nI1217 08:21:33.352116 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1217 08:21:33.352128 6165 handler.go:208] Removed *v1.Node event handler 2\\\\nI1217 08:21:33.352135 6165 handler.go:208] Removed *v1.Node event handler 7\\\\nI1217 08:21:33.352142 6165 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1217 08:21:33.352149 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1217 08:21:33.352268 6165 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352371 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352445 6165 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352725 6165 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.216450 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.216492 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.216503 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.216538 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.216553 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.224490 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.235735 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.244950 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/216f8233-5cc0-4733-b27e-69f9312d9a94-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.245011 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/216f8233-5cc0-4733-b27e-69f9312d9a94-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.245044 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n4cv\" (UniqueName: \"kubernetes.io/projected/216f8233-5cc0-4733-b27e-69f9312d9a94-kube-api-access-2n4cv\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.245075 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/216f8233-5cc0-4733-b27e-69f9312d9a94-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.246543 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/216f8233-5cc0-4733-b27e-69f9312d9a94-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.247627 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/216f8233-5cc0-4733-b27e-69f9312d9a94-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.247617 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.258699 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/216f8233-5cc0-4733-b27e-69f9312d9a94-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.263607 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n4cv\" (UniqueName: \"kubernetes.io/projected/216f8233-5cc0-4733-b27e-69f9312d9a94-kube-api-access-2n4cv\") pod \"ovnkube-control-plane-749d76644c-mw9pk\" (UID: \"216f8233-5cc0-4733-b27e-69f9312d9a94\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.264594 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.277622 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.289209 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.301072 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.312460 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.318571 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.318640 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.318650 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.318663 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.318695 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.322492 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.332332 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.343114 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.351302 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.359750 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.375642 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.385863 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.394712 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.401561 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.404467 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: W1217 08:21:36.414353 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod216f8233_5cc0_4733_b27e_69f9312d9a94.slice/crio-ccc738cee9c4bfe40158ff0880da119c52c7ac60844a8dbcdad876c96f7500f9 WatchSource:0}: Error finding container ccc738cee9c4bfe40158ff0880da119c52c7ac60844a8dbcdad876c96f7500f9: Status 404 returned error can't find the container with id ccc738cee9c4bfe40158ff0880da119c52c7ac60844a8dbcdad876c96f7500f9 Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.421224 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.421254 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.421296 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.421310 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.421320 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.426684 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7856659d189521a39f0069cd8718a6a02d9933a42588358df4424c3559e5ca47\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:33Z\\\",\\\"message\\\":\\\"event handler 7 for removal\\\\nI1217 08:21:33.352067 6165 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1217 08:21:33.352079 6165 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1217 08:21:33.352089 6165 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1217 08:21:33.352102 6165 factory.go:656] Stopping watch factory\\\\nI1217 08:21:33.352116 6165 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1217 08:21:33.352128 6165 handler.go:208] Removed *v1.Node event handler 2\\\\nI1217 08:21:33.352135 6165 handler.go:208] Removed *v1.Node event handler 7\\\\nI1217 08:21:33.352142 6165 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1217 08:21:33.352149 6165 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1217 08:21:33.352268 6165 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352371 6165 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352445 6165 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:33.352725 6165 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.437125 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.451014 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.462459 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.479785 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:36Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.524146 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.524184 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.524194 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.524208 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.524218 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.626983 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.627024 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.627034 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.627049 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.627060 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.729644 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.729684 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.729694 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.729710 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.729721 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.829865 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:36 crc kubenswrapper[4966]: E1217 08:21:36.830039 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.838172 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.838210 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.838220 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.838236 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.838247 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.941082 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.941132 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.941147 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.941168 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:36 crc kubenswrapper[4966]: I1217 08:21:36.941184 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:36Z","lastTransitionTime":"2025-12-17T08:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.043802 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.043913 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.043939 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.043970 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.043988 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.108415 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" event={"ID":"216f8233-5cc0-4733-b27e-69f9312d9a94","Type":"ContainerStarted","Data":"ccc738cee9c4bfe40158ff0880da119c52c7ac60844a8dbcdad876c96f7500f9"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.110183 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/1.log" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.113551 4966 scope.go:117] "RemoveContainer" containerID="f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c" Dec 17 08:21:37 crc kubenswrapper[4966]: E1217 08:21:37.113780 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.147332 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.148224 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.148266 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.148288 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.148320 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.148342 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.163283 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.177492 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.190190 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.210622 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.223308 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.237426 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.251064 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.251101 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.251111 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.251128 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.251139 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.251489 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.267570 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.278851 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.289244 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.296801 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.306478 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.317063 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.325388 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.335939 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.353405 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.353438 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.353450 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.353467 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.353481 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.456898 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.456928 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.456939 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.456954 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.456964 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.524650 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-nqmws"] Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.525446 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:37 crc kubenswrapper[4966]: E1217 08:21:37.525569 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.544499 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.559988 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.560015 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.560022 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.560035 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.560043 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.563606 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.576116 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.590550 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.607417 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.619178 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.629563 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.646861 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.656592 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.658761 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdrrh\" (UniqueName: \"kubernetes.io/projected/6cf617b8-bd09-49de-b188-9c86d5fafd57-kube-api-access-qdrrh\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.658801 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.662038 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.662096 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.662113 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.662140 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.662157 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.669237 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.682793 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.699894 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.710920 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.721849 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.732679 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.749294 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.759495 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdrrh\" (UniqueName: \"kubernetes.io/projected/6cf617b8-bd09-49de-b188-9c86d5fafd57-kube-api-access-qdrrh\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.759536 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:37 crc kubenswrapper[4966]: E1217 08:21:37.759648 4966 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:37 crc kubenswrapper[4966]: E1217 08:21:37.759699 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs podName:6cf617b8-bd09-49de-b188-9c86d5fafd57 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:38.259685816 +0000 UTC m=+33.804755758 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs") pod "network-metrics-daemon-nqmws" (UID: "6cf617b8-bd09-49de-b188-9c86d5fafd57") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.760387 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:37Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.764361 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.764397 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.764409 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.764425 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.764435 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.777527 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdrrh\" (UniqueName: \"kubernetes.io/projected/6cf617b8-bd09-49de-b188-9c86d5fafd57-kube-api-access-qdrrh\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.829630 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.829646 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:37 crc kubenswrapper[4966]: E1217 08:21:37.829752 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:37 crc kubenswrapper[4966]: E1217 08:21:37.829861 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.866682 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.866732 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.866749 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.866774 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.866793 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.969929 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.969976 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.969991 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.970011 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:37 crc kubenswrapper[4966]: I1217 08:21:37.970026 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:37Z","lastTransitionTime":"2025-12-17T08:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.073646 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.073716 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.073729 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.073746 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.073759 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.176260 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.176334 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.176356 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.176407 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.176433 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.265181 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.265310 4966 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.265359 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs podName:6cf617b8-bd09-49de-b188-9c86d5fafd57 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:39.265344659 +0000 UTC m=+34.810414601 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs") pod "network-metrics-daemon-nqmws" (UID: "6cf617b8-bd09-49de-b188-9c86d5fafd57") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.278894 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.278933 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.278945 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.278961 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.278972 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.381724 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.381776 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.381788 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.381807 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.381819 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.466319 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.466517 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:21:54.466491342 +0000 UTC m=+50.011561284 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.483660 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.483696 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.483706 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.483720 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.483730 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.567034 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.567116 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.567186 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.567249 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567382 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567449 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567476 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567510 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567376 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567550 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567571 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:54.567537987 +0000 UTC m=+50.112607969 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567579 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567615 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:54.567591408 +0000 UTC m=+50.112661380 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567660 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:54.567634029 +0000 UTC m=+50.112704031 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.567681 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.568011 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:54.567986228 +0000 UTC m=+50.113056290 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.586366 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.586452 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.586474 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.586499 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.586517 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.688448 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.688492 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.688503 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.688520 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.688531 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.791191 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.791244 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.791260 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.791284 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.791300 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.834629 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:38 crc kubenswrapper[4966]: E1217 08:21:38.834824 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.894190 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.894240 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.894253 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.894271 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.894283 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.997122 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.997162 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.997173 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.997210 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:38 crc kubenswrapper[4966]: I1217 08:21:38.997223 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:38Z","lastTransitionTime":"2025-12-17T08:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.100072 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.100356 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.100366 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.100381 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.100393 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:39Z","lastTransitionTime":"2025-12-17T08:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.119670 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" event={"ID":"216f8233-5cc0-4733-b27e-69f9312d9a94","Type":"ContainerStarted","Data":"9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.119710 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" event={"ID":"216f8233-5cc0-4733-b27e-69f9312d9a94","Type":"ContainerStarted","Data":"7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.137120 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.151478 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.160979 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.170663 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.181170 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.195634 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.203446 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.203488 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.203498 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.203515 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.203525 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:39Z","lastTransitionTime":"2025-12-17T08:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.209352 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.219686 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.249628 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.261554 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.272214 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.274558 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:39 crc kubenswrapper[4966]: E1217 08:21:39.274747 4966 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:39 crc kubenswrapper[4966]: E1217 08:21:39.274818 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs podName:6cf617b8-bd09-49de-b188-9c86d5fafd57 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:41.274796925 +0000 UTC m=+36.819866877 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs") pod "network-metrics-daemon-nqmws" (UID: "6cf617b8-bd09-49de-b188-9c86d5fafd57") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.284991 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.305410 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.305444 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.305455 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.305470 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.305481 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:39Z","lastTransitionTime":"2025-12-17T08:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.306545 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.316995 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.331126 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.343573 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.358186 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:39Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.407993 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.408058 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.408079 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.408103 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.408120 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:39Z","lastTransitionTime":"2025-12-17T08:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.510595 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.510697 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.510719 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.510782 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.510801 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:39Z","lastTransitionTime":"2025-12-17T08:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.612586 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.612643 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.612661 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.612685 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.612948 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:39Z","lastTransitionTime":"2025-12-17T08:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.716216 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.716308 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.716327 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.716350 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.716366 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:39Z","lastTransitionTime":"2025-12-17T08:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.818647 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.818695 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.818713 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.818734 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.818750 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:39Z","lastTransitionTime":"2025-12-17T08:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.830307 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.830343 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.830351 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:39 crc kubenswrapper[4966]: E1217 08:21:39.830481 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:39 crc kubenswrapper[4966]: E1217 08:21:39.830543 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:39 crc kubenswrapper[4966]: E1217 08:21:39.830653 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.921688 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.922146 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.922358 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.922560 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:39 crc kubenswrapper[4966]: I1217 08:21:39.922771 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:39Z","lastTransitionTime":"2025-12-17T08:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.026000 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.026074 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.026098 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.026129 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.026151 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.128805 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.128836 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.128845 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.128859 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.128888 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.231525 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.231564 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.231588 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.231607 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.231621 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.334171 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.334205 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.334214 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.334226 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.334234 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.436706 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.437035 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.437156 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.437241 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.437342 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.540124 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.540421 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.540507 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.540614 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.540720 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.644041 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.644113 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.644133 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.644167 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.644206 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.747460 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.747525 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.747542 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.747566 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.747585 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.829721 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:40 crc kubenswrapper[4966]: E1217 08:21:40.829933 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.849967 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.850045 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.850059 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.850080 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.850093 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.952754 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.952805 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.952817 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.952840 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:40 crc kubenswrapper[4966]: I1217 08:21:40.952852 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:40Z","lastTransitionTime":"2025-12-17T08:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.055506 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.055950 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.056131 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.056283 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.056465 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:41Z","lastTransitionTime":"2025-12-17T08:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.159927 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.160063 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.160083 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.160111 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.160128 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:41Z","lastTransitionTime":"2025-12-17T08:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.263463 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.263528 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.263548 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.263574 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.263591 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:41Z","lastTransitionTime":"2025-12-17T08:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.299436 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:41 crc kubenswrapper[4966]: E1217 08:21:41.299650 4966 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:41 crc kubenswrapper[4966]: E1217 08:21:41.299780 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs podName:6cf617b8-bd09-49de-b188-9c86d5fafd57 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:45.299744371 +0000 UTC m=+40.844814353 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs") pod "network-metrics-daemon-nqmws" (UID: "6cf617b8-bd09-49de-b188-9c86d5fafd57") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.367280 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.367361 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.367380 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.367408 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.367427 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:41Z","lastTransitionTime":"2025-12-17T08:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.470465 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.470499 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.470507 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.470520 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.470529 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:41Z","lastTransitionTime":"2025-12-17T08:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.573659 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.573731 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.573754 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.573786 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.573808 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:41Z","lastTransitionTime":"2025-12-17T08:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.611230 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.612247 4966 scope.go:117] "RemoveContainer" containerID="f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c" Dec 17 08:21:41 crc kubenswrapper[4966]: E1217 08:21:41.612428 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.755165 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.755224 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.755235 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.755266 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.755277 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:41Z","lastTransitionTime":"2025-12-17T08:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.829812 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.829930 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:41 crc kubenswrapper[4966]: E1217 08:21:41.829981 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:41 crc kubenswrapper[4966]: E1217 08:21:41.830125 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.829943 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:41 crc kubenswrapper[4966]: E1217 08:21:41.830256 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.857893 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.857953 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.857965 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.858002 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.858026 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:41Z","lastTransitionTime":"2025-12-17T08:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.960997 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.961056 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.961072 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.961096 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:41 crc kubenswrapper[4966]: I1217 08:21:41.961113 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:41Z","lastTransitionTime":"2025-12-17T08:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.063944 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.064024 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.064042 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.064068 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.064085 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.166703 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.166763 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.166792 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.166822 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.166843 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.269769 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.269839 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.269863 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.269935 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.269963 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.373058 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.373122 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.373161 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.373202 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.373242 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.477701 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.477768 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.477793 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.477823 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.477846 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.580207 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.580254 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.580265 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.580284 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.580296 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.667255 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.683339 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.683635 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.683804 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.684003 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.684172 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.687738 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.711041 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.733723 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.749769 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.766919 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.784267 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.787608 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.787641 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.787654 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.787674 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.787688 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.805294 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.820217 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.829496 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:42 crc kubenswrapper[4966]: E1217 08:21:42.829635 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.834784 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.856217 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.885616 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.890354 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.890550 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.890651 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.890741 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.890819 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.899812 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.911345 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.924052 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.943188 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.955051 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.966648 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:42Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.993470 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.993518 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.993538 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.993555 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:42 crc kubenswrapper[4966]: I1217 08:21:42.993570 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:42Z","lastTransitionTime":"2025-12-17T08:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.095530 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.095602 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.095620 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.095644 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.095695 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:43Z","lastTransitionTime":"2025-12-17T08:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.199466 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.199542 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.199559 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.199583 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.199600 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:43Z","lastTransitionTime":"2025-12-17T08:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.303190 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.303273 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.303292 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.303313 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.303326 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:43Z","lastTransitionTime":"2025-12-17T08:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.406347 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.406443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.406461 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.406486 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.406504 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:43Z","lastTransitionTime":"2025-12-17T08:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.508662 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.508690 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.508700 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.508713 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.508722 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:43Z","lastTransitionTime":"2025-12-17T08:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.611793 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.611834 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.611845 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.611862 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.611904 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:43Z","lastTransitionTime":"2025-12-17T08:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.714385 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.714612 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.714728 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.714844 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.714959 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:43Z","lastTransitionTime":"2025-12-17T08:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.818275 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.818307 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.818315 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.818330 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.818339 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:43Z","lastTransitionTime":"2025-12-17T08:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.829807 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.829916 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:43 crc kubenswrapper[4966]: E1217 08:21:43.829971 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:43 crc kubenswrapper[4966]: E1217 08:21:43.830124 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.830207 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:43 crc kubenswrapper[4966]: E1217 08:21:43.830323 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.920628 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.920659 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.920668 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.920682 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:43 crc kubenswrapper[4966]: I1217 08:21:43.920692 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:43Z","lastTransitionTime":"2025-12-17T08:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.022954 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.023006 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.023025 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.023048 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.023066 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.125857 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.125955 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.125970 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.125988 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.126000 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.228586 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.228630 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.228639 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.228658 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.228668 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.330506 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.330552 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.330562 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.330580 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.330591 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.432929 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.432968 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.432980 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.432998 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.433009 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.536127 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.536217 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.536229 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.536245 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.536257 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.638209 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.638270 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.638286 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.638650 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.638698 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.741238 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.741297 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.741309 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.741326 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.741337 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.829809 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:44 crc kubenswrapper[4966]: E1217 08:21:44.829994 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.844240 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.844538 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.844644 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.844723 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.844783 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.844262 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.853481 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.864399 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.876680 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.891226 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.906064 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.916557 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.926900 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.937968 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.955459 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.955491 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.955502 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.955518 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.955528 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:44Z","lastTransitionTime":"2025-12-17T08:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.965112 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:44 crc kubenswrapper[4966]: I1217 08:21:44.992490 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:44Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.020535 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:45Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.030689 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:45Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.041064 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:45Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.051453 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:45Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.058131 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.058153 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.058162 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.058175 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.058185 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.062769 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:45Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.075537 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:45Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.160508 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.160568 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.160591 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.160621 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.160642 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.263618 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.263682 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.263699 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.263767 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.263785 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.366304 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.366361 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.366384 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.366413 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.366435 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.383341 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:45 crc kubenswrapper[4966]: E1217 08:21:45.383498 4966 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:45 crc kubenswrapper[4966]: E1217 08:21:45.383604 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs podName:6cf617b8-bd09-49de-b188-9c86d5fafd57 nodeName:}" failed. No retries permitted until 2025-12-17 08:21:53.383574504 +0000 UTC m=+48.928644486 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs") pod "network-metrics-daemon-nqmws" (UID: "6cf617b8-bd09-49de-b188-9c86d5fafd57") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.469436 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.469609 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.469655 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.469713 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.469739 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.573729 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.573800 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.573821 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.573845 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.573863 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.677186 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.677253 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.677271 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.677302 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.677321 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.780037 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.780074 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.780086 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.780102 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.780112 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.829642 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.829679 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.829659 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:45 crc kubenswrapper[4966]: E1217 08:21:45.829761 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:45 crc kubenswrapper[4966]: E1217 08:21:45.829906 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:45 crc kubenswrapper[4966]: E1217 08:21:45.830060 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.882762 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.882809 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.882823 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.882839 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.882850 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.951150 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.951205 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.951223 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.951246 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.951261 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:45 crc kubenswrapper[4966]: E1217 08:21:45.972322 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:45Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.977940 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.978026 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.978043 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.978071 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:45 crc kubenswrapper[4966]: I1217 08:21:45.978090 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:45Z","lastTransitionTime":"2025-12-17T08:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: E1217 08:21:46.000054 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:45Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.004181 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.004265 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.004292 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.004323 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.004345 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: E1217 08:21:46.024489 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:46Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.028605 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.028660 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.028675 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.028695 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.028711 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: E1217 08:21:46.041486 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:46Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.044787 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.044839 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.044851 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.044866 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.044894 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: E1217 08:21:46.055121 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:46Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:46 crc kubenswrapper[4966]: E1217 08:21:46.055261 4966 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.056719 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.056756 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.056766 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.056784 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.056796 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.158679 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.158724 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.158734 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.158772 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.158790 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.261004 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.261051 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.261062 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.261082 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.261097 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.363047 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.363089 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.363104 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.363121 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.363133 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.464838 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.464888 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.464902 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.464921 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.464937 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.568156 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.568205 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.568217 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.568238 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.568247 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.672314 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.672388 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.672414 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.672446 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.672467 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.775237 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.775345 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.775370 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.775402 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.775423 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.830406 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:46 crc kubenswrapper[4966]: E1217 08:21:46.830648 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.878243 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.878297 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.878311 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.878331 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.878361 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.981265 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.981325 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.981337 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.981386 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:46 crc kubenswrapper[4966]: I1217 08:21:46.981398 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:46Z","lastTransitionTime":"2025-12-17T08:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.084983 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.085051 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.085069 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.085093 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.085110 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:47Z","lastTransitionTime":"2025-12-17T08:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.188282 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.188334 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.188352 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.188376 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.188393 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:47Z","lastTransitionTime":"2025-12-17T08:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.294177 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.294244 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.294261 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.294290 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.294311 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:47Z","lastTransitionTime":"2025-12-17T08:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.396980 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.397047 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.397069 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.397099 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.397117 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:47Z","lastTransitionTime":"2025-12-17T08:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.500278 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.500329 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.500346 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.500371 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.500389 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:47Z","lastTransitionTime":"2025-12-17T08:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.603411 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.603463 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.603475 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.603492 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.603507 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:47Z","lastTransitionTime":"2025-12-17T08:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.706097 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.706165 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.706212 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.706230 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.706241 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:47Z","lastTransitionTime":"2025-12-17T08:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.808556 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.808614 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.808626 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.808645 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.809353 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:47Z","lastTransitionTime":"2025-12-17T08:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.829827 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.829907 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:47 crc kubenswrapper[4966]: E1217 08:21:47.829952 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:47 crc kubenswrapper[4966]: E1217 08:21:47.830086 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.830177 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:47 crc kubenswrapper[4966]: E1217 08:21:47.830241 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.911737 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.911775 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.911786 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.911807 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:47 crc kubenswrapper[4966]: I1217 08:21:47.911821 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:47Z","lastTransitionTime":"2025-12-17T08:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.014178 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.014216 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.014227 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.014242 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.014253 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.117058 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.117104 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.117117 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.117136 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.117148 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.220045 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.220107 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.220119 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.220146 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.220161 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.323420 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.323477 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.323502 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.323533 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.323556 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.426115 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.426239 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.426259 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.426284 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.426303 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.530174 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.530251 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.530278 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.530308 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.530330 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.634438 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.634493 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.634516 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.634545 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.634566 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.737526 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.737597 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.737620 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.737651 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.737674 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.829506 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:48 crc kubenswrapper[4966]: E1217 08:21:48.829738 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.840105 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.840177 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.840199 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.840227 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.840252 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.944042 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.944173 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.944238 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.944272 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:48 crc kubenswrapper[4966]: I1217 08:21:48.944363 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:48Z","lastTransitionTime":"2025-12-17T08:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.046960 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.047001 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.047011 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.047028 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.047040 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.151557 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.151611 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.151630 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.151655 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.151673 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.258935 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.259002 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.259015 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.259037 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.259054 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.362580 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.362638 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.362656 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.362682 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.362700 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.465009 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.465048 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.465058 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.465075 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.465086 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.567237 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.567325 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.567349 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.567380 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.567401 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.670533 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.670582 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.670590 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.670606 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.670615 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.773254 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.773315 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.773338 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.773361 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.773374 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.830418 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.830507 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.830566 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:49 crc kubenswrapper[4966]: E1217 08:21:49.830867 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:49 crc kubenswrapper[4966]: E1217 08:21:49.830929 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:49 crc kubenswrapper[4966]: E1217 08:21:49.831010 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.876493 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.876539 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.876550 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.876568 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.876583 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.979381 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.979455 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.979477 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.979511 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:49 crc kubenswrapper[4966]: I1217 08:21:49.979531 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:49Z","lastTransitionTime":"2025-12-17T08:21:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.081553 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.081613 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.081626 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.081643 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.081655 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:50Z","lastTransitionTime":"2025-12-17T08:21:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.184348 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.184403 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.184411 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.184424 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.184433 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:50Z","lastTransitionTime":"2025-12-17T08:21:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.287558 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.287608 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.287622 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.287636 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.287647 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:50Z","lastTransitionTime":"2025-12-17T08:21:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.389580 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.389617 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.389625 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.389642 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.389651 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:50Z","lastTransitionTime":"2025-12-17T08:21:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.492512 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.492570 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.492586 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.492609 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.492627 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:50Z","lastTransitionTime":"2025-12-17T08:21:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.595207 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.595249 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.595261 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.595277 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.595288 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:50Z","lastTransitionTime":"2025-12-17T08:21:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.698514 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.698575 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.698587 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.698612 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.698626 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:50Z","lastTransitionTime":"2025-12-17T08:21:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.801991 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.802041 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.802060 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.802084 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.802101 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:50Z","lastTransitionTime":"2025-12-17T08:21:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.830587 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:50 crc kubenswrapper[4966]: E1217 08:21:50.830757 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.905947 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.905991 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.906001 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.906019 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:50 crc kubenswrapper[4966]: I1217 08:21:50.906030 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:50Z","lastTransitionTime":"2025-12-17T08:21:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.008151 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.008344 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.008365 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.008409 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.008424 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.111292 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.111329 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.111338 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.111356 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.111366 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.213646 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.213702 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.213718 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.213736 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.213746 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.316381 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.316444 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.316456 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.316472 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.316507 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.419596 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.419663 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.419687 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.419718 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.419739 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.522849 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.522945 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.522961 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.522979 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.522992 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.626209 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.626266 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.626282 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.626300 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.626318 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.728342 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.728405 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.728422 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.728448 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.728465 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.829749 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.829834 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:51 crc kubenswrapper[4966]: E1217 08:21:51.829998 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.830042 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:51 crc kubenswrapper[4966]: E1217 08:21:51.830277 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.831256 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.831280 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.831291 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.831418 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.831443 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:51 crc kubenswrapper[4966]: E1217 08:21:51.833085 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.934042 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.934100 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.934109 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.934123 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:51 crc kubenswrapper[4966]: I1217 08:21:51.934131 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:51Z","lastTransitionTime":"2025-12-17T08:21:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.036143 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.036169 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.036177 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.036195 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.036204 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.138408 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.138436 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.138444 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.138458 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.138467 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.242111 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.242156 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.242172 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.242196 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.242212 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.345465 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.345534 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.345555 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.345586 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.345608 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.448087 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.448127 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.448138 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.448154 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.448164 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.550487 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.550552 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.550571 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.550592 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.550607 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.653138 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.653184 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.653196 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.653214 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.653226 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.755184 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.755260 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.755273 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.755291 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.755307 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.830209 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:52 crc kubenswrapper[4966]: E1217 08:21:52.830410 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.858308 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.858348 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.858358 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.858373 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.858382 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.960887 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.960975 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.960992 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.961015 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:52 crc kubenswrapper[4966]: I1217 08:21:52.961031 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:52Z","lastTransitionTime":"2025-12-17T08:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.064148 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.064226 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.064244 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.064264 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.064280 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.166867 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.166937 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.166950 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.166973 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.166988 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.273676 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.273764 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.273808 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.273842 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.273866 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.377206 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.377244 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.377254 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.377270 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.377281 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.469933 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:53 crc kubenswrapper[4966]: E1217 08:21:53.470118 4966 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:53 crc kubenswrapper[4966]: E1217 08:21:53.470198 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs podName:6cf617b8-bd09-49de-b188-9c86d5fafd57 nodeName:}" failed. No retries permitted until 2025-12-17 08:22:09.470175451 +0000 UTC m=+65.015245393 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs") pod "network-metrics-daemon-nqmws" (UID: "6cf617b8-bd09-49de-b188-9c86d5fafd57") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.480093 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.480129 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.480142 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.480160 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.480171 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.583231 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.583293 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.583316 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.583342 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.583359 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.685800 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.685841 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.685850 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.685864 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.685892 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.788389 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.788438 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.788453 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.788476 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.788493 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.830077 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.830165 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.830187 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:53 crc kubenswrapper[4966]: E1217 08:21:53.830263 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:53 crc kubenswrapper[4966]: E1217 08:21:53.830418 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:53 crc kubenswrapper[4966]: E1217 08:21:53.830506 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.890948 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.891012 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.891024 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.891043 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.891056 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.993555 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.993610 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.993624 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.993646 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:53 crc kubenswrapper[4966]: I1217 08:21:53.993659 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:53Z","lastTransitionTime":"2025-12-17T08:21:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.096199 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.096262 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.096285 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.096315 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.096338 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:54Z","lastTransitionTime":"2025-12-17T08:21:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.199955 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.200029 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.200051 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.200078 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.200095 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:54Z","lastTransitionTime":"2025-12-17T08:21:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.302645 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.302700 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.302711 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.302729 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.302743 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:54Z","lastTransitionTime":"2025-12-17T08:21:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.406008 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.406062 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.406072 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.406088 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.406099 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:54Z","lastTransitionTime":"2025-12-17T08:21:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.482289 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.482531 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:22:26.482501684 +0000 UTC m=+82.027571626 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.508998 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.509072 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.509084 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.509132 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.509158 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:54Z","lastTransitionTime":"2025-12-17T08:21:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.583338 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.583392 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.583423 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.583453 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583496 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583556 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583572 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583577 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:22:26.583553749 +0000 UTC m=+82.128623711 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583585 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583622 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:22:26.58360911 +0000 UTC m=+82.128679062 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583628 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583657 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583671 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583680 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583709 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:22:26.583699793 +0000 UTC m=+82.128769745 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.583737 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:22:26.583717424 +0000 UTC m=+82.128787426 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.612807 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.612855 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.612866 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.612900 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.612912 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:54Z","lastTransitionTime":"2025-12-17T08:21:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.715063 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.715096 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.715107 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.715122 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.715135 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:54Z","lastTransitionTime":"2025-12-17T08:21:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.817749 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.817785 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.817793 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.817809 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.817818 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:54Z","lastTransitionTime":"2025-12-17T08:21:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.830122 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:54 crc kubenswrapper[4966]: E1217 08:21:54.830246 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.843290 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.856517 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.867749 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.881347 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.894028 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.908849 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.920925 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.921736 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.922150 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.922447 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.922850 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:54Z","lastTransitionTime":"2025-12-17T08:21:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.922211 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.934152 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.946459 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.959930 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.973368 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:54 crc kubenswrapper[4966]: I1217 08:21:54.983515 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.000349 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:54Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.010189 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:55Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.024573 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.024598 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.024632 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.024651 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.024661 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.031898 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:55Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.043843 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:55Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.063424 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:55Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.126924 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.126950 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.126958 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.126972 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.126983 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.229964 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.230026 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.230044 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.230067 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.230087 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.333541 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.333604 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.333626 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.333656 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.333678 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.435966 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.436018 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.436030 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.436047 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.436057 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.538196 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.538250 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.538264 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.538287 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.538299 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.640154 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.640181 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.640189 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.640202 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.640210 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.743240 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.743282 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.743297 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.743319 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.743333 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.830186 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.830231 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.830194 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:55 crc kubenswrapper[4966]: E1217 08:21:55.830350 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:55 crc kubenswrapper[4966]: E1217 08:21:55.830305 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:55 crc kubenswrapper[4966]: E1217 08:21:55.830464 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.831815 4966 scope.go:117] "RemoveContainer" containerID="f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.845196 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.845237 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.845249 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.845266 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.845278 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.947212 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.947478 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.947488 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.947502 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:55 crc kubenswrapper[4966]: I1217 08:21:55.947512 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:55Z","lastTransitionTime":"2025-12-17T08:21:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.050524 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.050567 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.050576 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.050592 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.050601 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.152569 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.152618 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.152629 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.152647 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.152658 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.178217 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/1.log" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.179911 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.180680 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.198131 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.211608 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.227704 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.250995 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.254113 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.254149 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.254161 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.254178 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.254190 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.265842 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.279780 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.290314 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.307261 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.316499 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.333651 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.352782 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.355927 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.355972 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.355984 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.356009 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.356021 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.367032 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.393618 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.404422 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.404470 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.404482 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.404498 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.404523 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.407070 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: E1217 08:21:56.416245 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.419538 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.419566 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.419578 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.419594 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.419604 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.425197 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: E1217 08:21:56.430524 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.433226 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.433256 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.433266 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.433281 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.433290 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.440790 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: E1217 08:21:56.443536 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.446215 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.446243 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.446251 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.446265 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.446274 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.461765 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: E1217 08:21:56.467477 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.470322 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.470432 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.470495 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.470574 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.470652 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: E1217 08:21:56.481561 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: E1217 08:21:56.481954 4966 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.483590 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.483695 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.483792 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.483864 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.483965 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.607639 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.607710 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.607734 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.607762 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.607786 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.676994 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.688289 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.694185 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.706929 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.709947 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.710058 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.710129 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.710202 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.710277 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.719826 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.740533 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.748481 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.758391 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.771298 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.780431 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.791375 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.802183 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.818035 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.818263 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.818518 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.818783 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.818946 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.825377 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.831090 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:56 crc kubenswrapper[4966]: E1217 08:21:56.831208 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.837399 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.847878 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.858594 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.885063 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.899153 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.931859 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.932926 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.933047 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.933143 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.933240 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:56 crc kubenswrapper[4966]: I1217 08:21:56.933326 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:56Z","lastTransitionTime":"2025-12-17T08:21:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.035822 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.035863 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.035904 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.035921 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.035933 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.138552 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.138629 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.138652 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.138684 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.138707 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.240976 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.241024 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.241038 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.241059 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.241074 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.342980 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.343013 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.343021 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.343037 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.343046 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.445509 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.445542 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.445550 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.445564 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.445573 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.548802 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.548920 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.548938 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.548960 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.548977 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.651011 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.651071 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.651084 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.651101 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.651114 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.753809 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.753851 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.753863 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.753914 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.753931 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.829802 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.829826 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:57 crc kubenswrapper[4966]: E1217 08:21:57.829932 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:57 crc kubenswrapper[4966]: E1217 08:21:57.830028 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.830446 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:57 crc kubenswrapper[4966]: E1217 08:21:57.830694 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.856318 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.856354 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.856365 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.856382 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.856396 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.958636 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.958680 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.958695 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.958717 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:57 crc kubenswrapper[4966]: I1217 08:21:57.958733 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:57Z","lastTransitionTime":"2025-12-17T08:21:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.061404 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.061448 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.061458 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.061474 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.061485 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.164179 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.164235 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.164247 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.164263 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.164278 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.189109 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/2.log" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.190003 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/1.log" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.192941 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461" exitCode=1 Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.192979 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.193021 4966 scope.go:117] "RemoveContainer" containerID="f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.194269 4966 scope.go:117] "RemoveContainer" containerID="2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461" Dec 17 08:21:58 crc kubenswrapper[4966]: E1217 08:21:58.194602 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.212542 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.223952 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.233214 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.243531 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.258491 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:57Z\\\",\\\"message\\\":\\\"_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1217 08:21:56.929996 6544 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1217 08:21:56.930024 6544 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1217 08:21:56.931578 6544 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.265739 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.265773 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.265784 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.265800 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.265810 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.267238 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.276543 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d221151-66d5-4d15-82d6-30df6d6e7ce7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.286750 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.296544 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.310052 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.324335 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.334135 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.343148 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.352142 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.367721 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.367781 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.367791 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.367804 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.367822 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.372020 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.381346 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.390659 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.402043 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:58Z is after 2025-08-24T17:21:41Z" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.470491 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.470518 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.470529 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.470553 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.470563 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.573373 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.573481 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.573504 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.573536 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.573559 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.676261 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.676344 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.676366 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.676397 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.676419 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.779407 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.779458 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.779475 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.779500 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.779518 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.829550 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:21:58 crc kubenswrapper[4966]: E1217 08:21:58.829735 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.882285 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.882337 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.882351 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.882371 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.882385 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.984410 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.984454 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.984466 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.984483 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:58 crc kubenswrapper[4966]: I1217 08:21:58.984497 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:58Z","lastTransitionTime":"2025-12-17T08:21:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.087006 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.087042 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.087052 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.087068 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.087079 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:59Z","lastTransitionTime":"2025-12-17T08:21:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.190109 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.190157 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.190168 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.190183 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.190192 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:59Z","lastTransitionTime":"2025-12-17T08:21:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.196986 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/2.log" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.292544 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.292589 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.292597 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.292613 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.292622 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:59Z","lastTransitionTime":"2025-12-17T08:21:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.395608 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.395650 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.395659 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.395673 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.395682 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:59Z","lastTransitionTime":"2025-12-17T08:21:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.497856 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.497989 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.498012 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.498043 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.498062 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:59Z","lastTransitionTime":"2025-12-17T08:21:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.601245 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.601297 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.601313 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.601336 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.601352 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:59Z","lastTransitionTime":"2025-12-17T08:21:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.704094 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.704153 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.704166 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.704183 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.704193 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:59Z","lastTransitionTime":"2025-12-17T08:21:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.806234 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.806287 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.806301 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.806324 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.806339 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:59Z","lastTransitionTime":"2025-12-17T08:21:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.829746 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.829758 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.829763 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:21:59 crc kubenswrapper[4966]: E1217 08:21:59.829964 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:21:59 crc kubenswrapper[4966]: E1217 08:21:59.830206 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:21:59 crc kubenswrapper[4966]: E1217 08:21:59.830329 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.909508 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.909580 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.909606 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.909697 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:21:59 crc kubenswrapper[4966]: I1217 08:21:59.909718 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:21:59Z","lastTransitionTime":"2025-12-17T08:21:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.012468 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.012543 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.012565 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.012594 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.012614 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.115332 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.115366 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.115375 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.115389 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.115398 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.218756 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.218834 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.218857 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.218920 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.218945 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.322398 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.322460 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.322480 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.322504 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.322521 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.425603 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.425716 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.425740 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.425775 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.425799 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.528920 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.528956 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.528965 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.529002 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.529013 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.631713 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.631762 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.631775 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.631791 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.631803 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.734617 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.734681 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.734700 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.734725 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.734794 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.830007 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:00 crc kubenswrapper[4966]: E1217 08:22:00.830195 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.838243 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.838303 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.838313 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.838329 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.838339 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.941495 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.941574 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.941599 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.941632 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:00 crc kubenswrapper[4966]: I1217 08:22:00.941655 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:00Z","lastTransitionTime":"2025-12-17T08:22:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.044650 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.044687 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.044744 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.044768 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.044779 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.147542 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.147613 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.147638 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.147670 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.147693 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.250766 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.250830 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.250852 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.250919 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.250945 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.353693 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.353749 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.353767 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.353785 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.353793 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.455822 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.455897 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.455914 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.455935 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.455951 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.558631 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.558697 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.558708 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.558725 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.558736 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.661465 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.661543 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.661559 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.661585 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.661598 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.763992 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.764027 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.764037 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.764052 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.764060 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.829632 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.829680 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:01 crc kubenswrapper[4966]: E1217 08:22:01.829904 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.829632 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:01 crc kubenswrapper[4966]: E1217 08:22:01.829774 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:01 crc kubenswrapper[4966]: E1217 08:22:01.829967 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.867022 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.867080 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.867097 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.867117 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.867131 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.969417 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.969469 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.969478 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.969492 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:01 crc kubenswrapper[4966]: I1217 08:22:01.969500 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:01Z","lastTransitionTime":"2025-12-17T08:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.073149 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.073243 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.073260 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.073285 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.073301 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.176268 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.176313 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.176326 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.176343 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.176354 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.279031 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.279094 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.279111 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.279138 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.279157 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.381233 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.381310 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.381324 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.381342 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.381356 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.483380 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.483436 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.483447 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.483463 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.483472 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.586068 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.586152 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.586169 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.586194 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.586208 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.688046 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.688100 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.688110 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.688125 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.688136 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.791278 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.791322 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.791335 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.791351 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.791364 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.829707 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:02 crc kubenswrapper[4966]: E1217 08:22:02.829930 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.894130 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.894232 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.894255 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.894287 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.894311 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.996465 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.996504 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.996516 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.996533 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:02 crc kubenswrapper[4966]: I1217 08:22:02.996545 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:02Z","lastTransitionTime":"2025-12-17T08:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.099530 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.099598 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.099617 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.099642 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.099661 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:03Z","lastTransitionTime":"2025-12-17T08:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.203174 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.203234 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.203247 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.203268 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.203279 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:03Z","lastTransitionTime":"2025-12-17T08:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.306149 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.306204 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.306226 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.306247 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.306262 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:03Z","lastTransitionTime":"2025-12-17T08:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.409031 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.409070 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.409082 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.409099 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.409110 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:03Z","lastTransitionTime":"2025-12-17T08:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.512161 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.512206 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.512214 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.512230 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.512240 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:03Z","lastTransitionTime":"2025-12-17T08:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.615027 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.615074 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.615087 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.615104 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.615114 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:03Z","lastTransitionTime":"2025-12-17T08:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.717320 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.717360 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.717371 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.717387 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.717398 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:03Z","lastTransitionTime":"2025-12-17T08:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.820317 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.820445 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.820468 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.820500 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.820524 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:03Z","lastTransitionTime":"2025-12-17T08:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.829743 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.829777 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.829743 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:03 crc kubenswrapper[4966]: E1217 08:22:03.829926 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:03 crc kubenswrapper[4966]: E1217 08:22:03.830052 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:03 crc kubenswrapper[4966]: E1217 08:22:03.830274 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.924122 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.924161 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.924172 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.924231 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:03 crc kubenswrapper[4966]: I1217 08:22:03.924248 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:03Z","lastTransitionTime":"2025-12-17T08:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.027584 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.027710 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.027784 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.027817 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.027945 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.130788 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.130836 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.130848 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.130865 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.130896 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.233273 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.233334 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.233342 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.233355 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.233389 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.335334 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.335400 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.335408 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.335421 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.335429 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.437001 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.437048 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.437058 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.437070 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.437078 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.539948 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.540007 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.540024 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.540049 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.540065 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.643087 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.643171 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.643196 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.643221 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.643237 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.745492 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.745525 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.745533 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.745547 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.745555 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.829681 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:04 crc kubenswrapper[4966]: E1217 08:22:04.829936 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.847732 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.847772 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.847905 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.847917 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.847937 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.847950 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.865261 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:57Z\\\",\\\"message\\\":\\\"_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1217 08:21:56.929996 6544 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1217 08:21:56.930024 6544 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1217 08:21:56.931578 6544 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.874065 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.884231 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d221151-66d5-4d15-82d6-30df6d6e7ce7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.904743 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.916144 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.926446 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.937937 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.950263 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.950306 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.950315 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.950332 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.950341 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:04Z","lastTransitionTime":"2025-12-17T08:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.950456 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.962770 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.974970 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.984286 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:04 crc kubenswrapper[4966]: I1217 08:22:04.993561 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:04Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.004639 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:05Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.016347 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:05Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.026012 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:05Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.036645 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:05Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.048631 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:05Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.051972 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.052009 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.052017 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.052048 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.052057 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.154755 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.154805 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.154818 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.154834 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.154848 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.257508 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.257550 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.257561 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.257578 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.257590 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.360219 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.360259 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.360268 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.360284 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.360294 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.462907 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.462955 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.462970 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.462990 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.463005 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.565992 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.566042 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.566050 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.566066 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.566077 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.668157 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.668502 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.668519 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.668542 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.668559 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.794331 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.794365 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.794375 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.794390 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.794400 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.830202 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:05 crc kubenswrapper[4966]: E1217 08:22:05.830311 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.830471 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:05 crc kubenswrapper[4966]: E1217 08:22:05.830519 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.830814 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:05 crc kubenswrapper[4966]: E1217 08:22:05.830902 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.896280 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.896318 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.896326 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.896341 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.896352 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.999342 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.999390 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.999404 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.999422 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:05 crc kubenswrapper[4966]: I1217 08:22:05.999435 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:05Z","lastTransitionTime":"2025-12-17T08:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.102310 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.102373 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.102388 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.102406 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.102417 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.205720 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.205948 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.206035 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.206106 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.206167 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.308820 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.309249 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.309358 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.309435 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.309510 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.412925 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.412985 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.413004 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.413031 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.413049 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.515282 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.515324 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.515332 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.515346 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.515356 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.618563 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.618625 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.618638 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.618654 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.618667 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.637743 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.637782 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.637799 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.637856 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.637898 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: E1217 08:22:06.651250 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:06Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.655742 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.655821 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.655836 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.655856 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.655896 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: E1217 08:22:06.671065 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:06Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.674766 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.675070 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.675174 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.675294 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.675428 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: E1217 08:22:06.689441 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:06Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.693534 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.693574 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.693589 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.693609 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.693623 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: E1217 08:22:06.708916 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:06Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.712838 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.712903 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.712922 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.712943 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.712995 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: E1217 08:22:06.725188 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:06Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:06 crc kubenswrapper[4966]: E1217 08:22:06.725334 4966 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.726926 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.726956 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.726966 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.726982 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.726994 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.828959 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.828988 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.828996 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.829010 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.829021 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.829364 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:06 crc kubenswrapper[4966]: E1217 08:22:06.829461 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.931375 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.931431 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.931450 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.931473 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:06 crc kubenswrapper[4966]: I1217 08:22:06.931490 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:06Z","lastTransitionTime":"2025-12-17T08:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.034088 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.034151 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.034165 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.034181 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.034194 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.136813 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.136914 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.136927 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.136944 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.136956 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.239499 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.239540 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.239555 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.239576 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.239592 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.342850 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.342942 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.342955 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.342975 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.342989 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.445661 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.445701 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.445711 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.445727 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.445739 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.547413 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.547441 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.547448 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.547461 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.547469 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.649775 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.649815 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.649826 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.649844 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.649855 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.752293 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.752344 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.752360 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.752376 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.752387 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.834692 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.834851 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:07 crc kubenswrapper[4966]: E1217 08:22:07.835089 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.835140 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:07 crc kubenswrapper[4966]: E1217 08:22:07.835321 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.835427 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:07 crc kubenswrapper[4966]: E1217 08:22:07.835543 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:07 crc kubenswrapper[4966]: E1217 08:22:07.835650 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.854895 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.854947 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.854961 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.854983 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.854997 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.957579 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.957614 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.957623 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.957638 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:07 crc kubenswrapper[4966]: I1217 08:22:07.957647 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:07Z","lastTransitionTime":"2025-12-17T08:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.059522 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.060043 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.060153 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.060244 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.060433 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.163136 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.163170 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.163180 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.163195 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.163207 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.265434 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.265473 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.265482 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.265496 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.265504 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.367413 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.367495 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.367505 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.367520 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.367530 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.469471 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.469728 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.469827 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.469936 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.470025 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.571932 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.571960 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.571969 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.571983 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.571992 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.673790 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.674030 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.674099 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.674165 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.674252 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.776849 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.776924 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.776941 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.776963 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.776979 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.878712 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.878990 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.879077 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.879161 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.879235 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.982398 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.982456 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.982474 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.982499 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:08 crc kubenswrapper[4966]: I1217 08:22:08.982517 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:08Z","lastTransitionTime":"2025-12-17T08:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.084920 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.084953 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.084962 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.084975 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.084984 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:09Z","lastTransitionTime":"2025-12-17T08:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.186634 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.186858 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.186961 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.187039 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.187100 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:09Z","lastTransitionTime":"2025-12-17T08:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.289039 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.289077 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.289087 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.289103 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.289113 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:09Z","lastTransitionTime":"2025-12-17T08:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.392087 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.392121 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.392132 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.392146 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.392155 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:09Z","lastTransitionTime":"2025-12-17T08:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.494710 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.494739 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.494748 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.494762 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.494771 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:09Z","lastTransitionTime":"2025-12-17T08:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.552705 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:09 crc kubenswrapper[4966]: E1217 08:22:09.552936 4966 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:22:09 crc kubenswrapper[4966]: E1217 08:22:09.553067 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs podName:6cf617b8-bd09-49de-b188-9c86d5fafd57 nodeName:}" failed. No retries permitted until 2025-12-17 08:22:41.553031949 +0000 UTC m=+97.098101921 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs") pod "network-metrics-daemon-nqmws" (UID: "6cf617b8-bd09-49de-b188-9c86d5fafd57") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.597485 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.597535 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.597551 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.597569 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.597582 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:09Z","lastTransitionTime":"2025-12-17T08:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.700618 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.700658 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.700666 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.700682 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.700696 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:09Z","lastTransitionTime":"2025-12-17T08:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.802594 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.802631 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.802640 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.802655 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.802663 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:09Z","lastTransitionTime":"2025-12-17T08:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.829772 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.829821 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.829774 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:09 crc kubenswrapper[4966]: E1217 08:22:09.829894 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.829780 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:09 crc kubenswrapper[4966]: E1217 08:22:09.829984 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:09 crc kubenswrapper[4966]: E1217 08:22:09.830020 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:09 crc kubenswrapper[4966]: E1217 08:22:09.830090 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.904918 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.904956 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.904965 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.904979 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:09 crc kubenswrapper[4966]: I1217 08:22:09.904988 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:09Z","lastTransitionTime":"2025-12-17T08:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.006630 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.006663 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.006671 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.006689 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.006699 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.109160 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.109673 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.109774 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.109842 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.109936 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.212119 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.212151 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.212160 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.212176 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.212184 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.313825 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.313861 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.313890 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.313918 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.313928 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.420479 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.420529 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.420541 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.420566 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.420582 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.522620 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.522653 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.522662 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.522676 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.522686 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.624816 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.624897 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.624909 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.624926 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.624936 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.727391 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.727435 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.727444 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.727459 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.727470 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.829903 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.829955 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.829964 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.829979 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.829988 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.932146 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.932182 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.932194 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.932210 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:10 crc kubenswrapper[4966]: I1217 08:22:10.932220 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:10Z","lastTransitionTime":"2025-12-17T08:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.033857 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.033911 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.033922 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.033937 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.033948 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.136356 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.136415 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.136426 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.136443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.136456 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.238723 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.238755 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.238763 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.238796 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.238804 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.340828 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.340884 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.340897 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.340914 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.340926 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.442917 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.442973 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.442985 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.443000 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.443012 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.544886 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.544928 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.544939 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.544957 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.544966 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.647361 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.647396 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.647406 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.647419 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.647428 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.749353 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.749386 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.749395 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.749409 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.749420 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.829936 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:11 crc kubenswrapper[4966]: E1217 08:22:11.830091 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.830110 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.830133 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.830152 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:11 crc kubenswrapper[4966]: E1217 08:22:11.830217 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:11 crc kubenswrapper[4966]: E1217 08:22:11.830305 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:11 crc kubenswrapper[4966]: E1217 08:22:11.830357 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.851830 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.851911 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.851924 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.851941 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.851953 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.953786 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.953843 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.953858 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.953904 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:11 crc kubenswrapper[4966]: I1217 08:22:11.953917 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:11Z","lastTransitionTime":"2025-12-17T08:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.055776 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.055814 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.055825 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.055842 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.055915 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.158477 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.158568 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.158587 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.158610 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.158627 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.240669 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/0.log" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.240766 4966 generic.go:334] "Generic (PLEG): container finished" podID="ac204c30-92c9-4b1d-9cda-bef16b7a0008" containerID="6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74" exitCode=1 Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.240838 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqk4j" event={"ID":"ac204c30-92c9-4b1d-9cda-bef16b7a0008","Type":"ContainerDied","Data":"6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.241744 4966 scope.go:117] "RemoveContainer" containerID="6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.253667 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.260713 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.260778 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.260795 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.260843 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.260861 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.267997 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.280559 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.292203 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.308416 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.324330 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.334916 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.346480 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.362266 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:22:11Z\\\",\\\"message\\\":\\\"2025-12-17T08:21:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1\\\\n2025-12-17T08:21:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1 to /host/opt/cni/bin/\\\\n2025-12-17T08:21:26Z [verbose] multus-daemon started\\\\n2025-12-17T08:21:26Z [verbose] Readiness Indicator file check\\\\n2025-12-17T08:22:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.363653 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.363674 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.363682 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.363695 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.363704 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.375111 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.398687 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f31e6e4786dfa3a60877dcddd48d83c99bb8b720a869eb22e5c25aa6ae087f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:35Z\\\",\\\"message\\\":\\\" 6285 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989393 6285 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989230 6285 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-image-registry/image-registry for endpointslice openshift-image-registry/image-registry-2vl4t as it is not a known egress service\\\\nI1217 08:21:34.989447 6285 egressservice_zone_node.go:110] Processing sync for Egress Service node crc\\\\nI1217 08:21:34.989456 6285 egressservice_zone_node.go:113] Finished syncing Egress Service node crc: 11.851µs\\\\nI1217 08:21:34.989561 6285 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989778 6285 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1217 08:21:34.989838 6285 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1217 08:21:34.990159 6285 ovnkube.go:599] Stopped ovnkube\\\\nI1217 08:21:34.990192 6285 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1217 08:21:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:57Z\\\",\\\"message\\\":\\\"_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1217 08:21:56.929996 6544 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1217 08:21:56.930024 6544 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1217 08:21:56.931578 6544 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.455939 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.466340 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.466375 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.466385 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.466398 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.466406 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.473366 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d221151-66d5-4d15-82d6-30df6d6e7ce7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.490166 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.500743 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.510159 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.520079 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.532007 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.568404 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.568443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.568454 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.568470 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.568480 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.670192 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.670231 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.670242 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.670258 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.670269 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.772683 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.772718 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.772726 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.772740 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.772749 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.831579 4966 scope.go:117] "RemoveContainer" containerID="2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461" Dec 17 08:22:12 crc kubenswrapper[4966]: E1217 08:22:12.831866 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.845545 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.858509 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.871096 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.874951 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.874975 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.874984 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.875000 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.875008 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.884338 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:22:11Z\\\",\\\"message\\\":\\\"2025-12-17T08:21:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1\\\\n2025-12-17T08:21:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1 to /host/opt/cni/bin/\\\\n2025-12-17T08:21:26Z [verbose] multus-daemon started\\\\n2025-12-17T08:21:26Z [verbose] Readiness Indicator file check\\\\n2025-12-17T08:22:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.898022 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.910182 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.925117 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.937640 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:12Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.977397 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.977436 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.977444 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.977458 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:12 crc kubenswrapper[4966]: I1217 08:22:12.977467 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:12Z","lastTransitionTime":"2025-12-17T08:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.032167 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.044209 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.054503 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.064107 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.079358 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.079568 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.079645 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.079712 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.079768 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:13Z","lastTransitionTime":"2025-12-17T08:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.080826 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:57Z\\\",\\\"message\\\":\\\"_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1217 08:21:56.929996 6544 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1217 08:21:56.930024 6544 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1217 08:21:56.931578 6544 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.090501 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.102411 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d221151-66d5-4d15-82d6-30df6d6e7ce7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.127527 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.141356 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.155301 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.182117 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.182153 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.182163 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.182177 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.182187 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:13Z","lastTransitionTime":"2025-12-17T08:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.245437 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/0.log" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.245488 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqk4j" event={"ID":"ac204c30-92c9-4b1d-9cda-bef16b7a0008","Type":"ContainerStarted","Data":"2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.259113 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.272036 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.280448 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.284355 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.284614 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.284692 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.284790 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.284859 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:13Z","lastTransitionTime":"2025-12-17T08:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.291189 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.301966 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:22:11Z\\\",\\\"message\\\":\\\"2025-12-17T08:21:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1\\\\n2025-12-17T08:21:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1 to /host/opt/cni/bin/\\\\n2025-12-17T08:21:26Z [verbose] multus-daemon started\\\\n2025-12-17T08:21:26Z [verbose] Readiness Indicator file check\\\\n2025-12-17T08:22:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:22:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.310230 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.319475 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.329671 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.345263 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d221151-66d5-4d15-82d6-30df6d6e7ce7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.370155 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.387814 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.388046 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.388059 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.388068 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.388081 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.388090 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:13Z","lastTransitionTime":"2025-12-17T08:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.400285 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.414722 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.433728 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:57Z\\\",\\\"message\\\":\\\"_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1217 08:21:56.929996 6544 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1217 08:21:56.930024 6544 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1217 08:21:56.931578 6544 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.444769 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.457978 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.469837 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.484939 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:13Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.490276 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.490346 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.490355 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.490370 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.490379 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:13Z","lastTransitionTime":"2025-12-17T08:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.592632 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.592674 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.592686 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.592705 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.592717 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:13Z","lastTransitionTime":"2025-12-17T08:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.694489 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.694535 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.694544 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.694558 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.694566 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:13Z","lastTransitionTime":"2025-12-17T08:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.796863 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.796928 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.796940 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.796958 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.796970 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:13Z","lastTransitionTime":"2025-12-17T08:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.830215 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.830251 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.830268 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.830277 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:13 crc kubenswrapper[4966]: E1217 08:22:13.830545 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:13 crc kubenswrapper[4966]: E1217 08:22:13.830640 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:13 crc kubenswrapper[4966]: E1217 08:22:13.830702 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:13 crc kubenswrapper[4966]: E1217 08:22:13.830752 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.899698 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.899730 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.899738 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.899753 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:13 crc kubenswrapper[4966]: I1217 08:22:13.899762 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:13Z","lastTransitionTime":"2025-12-17T08:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.001861 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.001914 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.001922 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.001962 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.001973 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.104113 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.104153 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.104165 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.104179 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.104188 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.206533 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.206560 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.206569 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.206581 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.206589 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.309003 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.309031 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.309040 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.309052 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.309060 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.410668 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.410699 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.410707 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.410719 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.410729 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.512443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.512496 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.512504 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.512518 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.512527 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.614508 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.614562 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.614572 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.614588 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.614599 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.717044 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.717078 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.717114 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.717134 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.717145 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.819224 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.819519 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.819646 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.819744 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.819826 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.846344 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.858701 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.870788 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.882198 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.896760 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.908901 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.919971 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:22:11Z\\\",\\\"message\\\":\\\"2025-12-17T08:21:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1\\\\n2025-12-17T08:21:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1 to /host/opt/cni/bin/\\\\n2025-12-17T08:21:26Z [verbose] multus-daemon started\\\\n2025-12-17T08:21:26Z [verbose] Readiness Indicator file check\\\\n2025-12-17T08:22:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:22:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.921825 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.921852 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.921862 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.921898 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.921910 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:14Z","lastTransitionTime":"2025-12-17T08:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.929345 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.939248 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.949636 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.960543 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d221151-66d5-4d15-82d6-30df6d6e7ce7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.980985 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:14 crc kubenswrapper[4966]: I1217 08:22:14.991437 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:14Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.002718 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:15Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.015345 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:15Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.023787 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.023828 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.023839 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.023856 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.023879 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.034986 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:57Z\\\",\\\"message\\\":\\\"_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1217 08:21:56.929996 6544 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1217 08:21:56.930024 6544 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1217 08:21:56.931578 6544 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:15Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.046599 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:15Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.056831 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:15Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.125256 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.125303 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.125317 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.125336 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.125348 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.227725 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.227767 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.227777 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.227791 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.227801 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.330328 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.330364 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.330376 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.330392 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.330402 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.432622 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.432669 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.432680 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.432696 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.432707 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.534761 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.534803 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.534814 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.534833 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.534846 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.637170 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.637206 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.637216 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.637232 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.637241 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.739461 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.739528 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.739540 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.739558 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.739571 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.829661 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.829675 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.829707 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.829685 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:15 crc kubenswrapper[4966]: E1217 08:22:15.829804 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:15 crc kubenswrapper[4966]: E1217 08:22:15.829919 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:15 crc kubenswrapper[4966]: E1217 08:22:15.829997 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:15 crc kubenswrapper[4966]: E1217 08:22:15.830128 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.842633 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.842667 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.842680 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.842696 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.842706 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.944503 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.944536 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.944547 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.944562 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:15 crc kubenswrapper[4966]: I1217 08:22:15.944573 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:15Z","lastTransitionTime":"2025-12-17T08:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.046395 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.046473 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.046496 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.046528 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.046551 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.149016 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.149047 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.149055 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.149069 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.149093 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.252443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.252480 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.252490 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.252507 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.252521 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.355272 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.355313 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.355325 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.355343 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.355357 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.457463 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.457497 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.457509 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.457523 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.457533 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.559382 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.559420 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.559432 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.559446 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.559457 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.662297 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.662331 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.662443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.662471 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.662484 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.765157 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.765190 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.765201 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.765216 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.765227 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.867775 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.867833 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.867847 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.867864 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.867904 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.970572 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.970613 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.970624 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.970640 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:16 crc kubenswrapper[4966]: I1217 08:22:16.970651 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:16Z","lastTransitionTime":"2025-12-17T08:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.073189 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.073234 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.073245 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.073261 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.073275 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.078335 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.078390 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.078403 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.078414 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.078425 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.090662 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:17Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.094034 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.094061 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.094070 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.094086 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.094096 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.105351 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:17Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.108333 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.108360 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.108370 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.108383 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.108392 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.118665 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:17Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.121353 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.121386 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.121396 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.121411 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.121420 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.131550 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:17Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.134722 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.134759 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.134769 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.134784 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.134793 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.146664 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:17Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.146778 4966 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.174955 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.174983 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.174991 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.175004 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.175013 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.276763 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.276804 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.276815 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.276830 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.276842 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.379270 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.379316 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.379328 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.379342 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.379352 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.483121 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.483145 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.483154 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.483167 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.483175 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.587317 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.587366 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.587382 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.587404 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.587420 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.689328 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.689368 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.689380 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.689395 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.689406 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.792244 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.792288 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.792300 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.792319 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.792336 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.829813 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.829965 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.829982 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.830139 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.830246 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.830242 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.830347 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:17 crc kubenswrapper[4966]: E1217 08:22:17.830395 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.894632 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.894669 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.894681 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.894699 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.894713 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.997167 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.997232 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.997250 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.997278 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:17 crc kubenswrapper[4966]: I1217 08:22:17.997299 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:17Z","lastTransitionTime":"2025-12-17T08:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.099273 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.099545 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.099660 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.099770 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.099894 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:18Z","lastTransitionTime":"2025-12-17T08:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.202157 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.202190 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.202202 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.202221 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.202232 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:18Z","lastTransitionTime":"2025-12-17T08:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.305614 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.306620 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.306801 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.307034 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.307191 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:18Z","lastTransitionTime":"2025-12-17T08:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.410126 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.410354 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.410482 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.410589 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.410714 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:18Z","lastTransitionTime":"2025-12-17T08:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.512733 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.513008 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.513098 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.513177 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.513243 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:18Z","lastTransitionTime":"2025-12-17T08:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.615116 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.615415 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.615498 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.615569 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.615636 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:18Z","lastTransitionTime":"2025-12-17T08:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.718753 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.718831 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.718855 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.718920 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.718944 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:18Z","lastTransitionTime":"2025-12-17T08:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.821781 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.821830 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.821845 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.821866 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.821916 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:18Z","lastTransitionTime":"2025-12-17T08:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.925470 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.925772 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.925940 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.926094 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:18 crc kubenswrapper[4966]: I1217 08:22:18.926207 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:18Z","lastTransitionTime":"2025-12-17T08:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.034228 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.034282 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.034305 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.034335 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.034358 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.137653 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.138072 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.138358 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.138551 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.138734 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.242293 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.242633 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.242820 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.243009 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.243211 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.345807 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.345838 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.345848 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.345862 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.345894 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.448411 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.448470 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.448500 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.448519 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.448531 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.551030 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.551061 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.551069 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.551082 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.551092 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.653656 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.654093 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.654236 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.654423 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.654581 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.757534 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.758006 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.758127 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.758230 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.758330 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.830249 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.830327 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:19 crc kubenswrapper[4966]: E1217 08:22:19.830684 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.830345 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:19 crc kubenswrapper[4966]: E1217 08:22:19.830823 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.830346 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:19 crc kubenswrapper[4966]: E1217 08:22:19.830834 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:19 crc kubenswrapper[4966]: E1217 08:22:19.831068 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.861241 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.861499 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.861577 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.861654 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.861720 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.964677 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.964747 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.964779 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.964809 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:19 crc kubenswrapper[4966]: I1217 08:22:19.964830 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:19Z","lastTransitionTime":"2025-12-17T08:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.067057 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.067118 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.067139 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.067162 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.067178 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.169641 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.169689 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.169703 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.169727 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.169742 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.273322 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.273644 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.273794 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.273956 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.274072 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.375931 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.375984 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.376001 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.376022 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.376037 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.478394 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.478440 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.478448 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.478463 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.478472 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.582016 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.582087 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.582106 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.582131 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.582149 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.685422 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.685486 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.685498 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.685515 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.685527 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.789294 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.789350 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.789367 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.789392 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.789410 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.892944 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.892991 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.893005 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.893026 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.893040 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.997467 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.997648 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.997866 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.997941 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:20 crc kubenswrapper[4966]: I1217 08:22:20.997979 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:20Z","lastTransitionTime":"2025-12-17T08:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.101653 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.101704 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.101715 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.101734 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.101746 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:21Z","lastTransitionTime":"2025-12-17T08:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.204838 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.204918 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.204934 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.204953 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.205318 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:21Z","lastTransitionTime":"2025-12-17T08:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.307492 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.307529 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.307539 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.307555 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.307565 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:21Z","lastTransitionTime":"2025-12-17T08:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.410227 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.410716 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.410829 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.410942 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.411048 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:21Z","lastTransitionTime":"2025-12-17T08:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.513671 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.513708 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.513719 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.513736 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.513747 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:21Z","lastTransitionTime":"2025-12-17T08:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.616472 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.617060 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.617385 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.617647 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.617971 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:21Z","lastTransitionTime":"2025-12-17T08:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.720830 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.720891 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.720907 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.720929 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.720945 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:21Z","lastTransitionTime":"2025-12-17T08:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.823352 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.823656 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.823736 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.823816 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.823928 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:21Z","lastTransitionTime":"2025-12-17T08:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.829686 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.829771 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.829993 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.830188 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:21 crc kubenswrapper[4966]: E1217 08:22:21.830183 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:21 crc kubenswrapper[4966]: E1217 08:22:21.830353 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:21 crc kubenswrapper[4966]: E1217 08:22:21.830629 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:21 crc kubenswrapper[4966]: E1217 08:22:21.830540 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.926299 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.926347 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.926358 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.926374 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:21 crc kubenswrapper[4966]: I1217 08:22:21.926386 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:21Z","lastTransitionTime":"2025-12-17T08:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.029434 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.029515 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.029533 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.029581 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.029594 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.132425 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.132575 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.132592 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.132615 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.132631 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.234826 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.234920 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.234938 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.234962 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.234978 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.337594 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.337675 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.337694 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.337721 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.337741 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.441008 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.441913 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.442162 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.442331 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.442478 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.546687 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.546742 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.546758 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.546780 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.546798 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.648686 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.648730 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.648745 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.648765 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.648779 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.751354 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.751394 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.751405 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.751421 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.751432 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.853719 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.853778 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.853800 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.853828 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.853850 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.956628 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.956697 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.956711 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.956729 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:22 crc kubenswrapper[4966]: I1217 08:22:22.956776 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:22Z","lastTransitionTime":"2025-12-17T08:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.060340 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.060413 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.060430 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.060456 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.060474 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.163811 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.163913 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.163939 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.163968 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.163989 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.266959 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.267006 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.267017 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.267036 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.267047 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.370557 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.370611 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.370635 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.370664 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.370690 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.474149 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.474203 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.474220 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.474242 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.474259 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.577535 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.577594 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.577613 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.577637 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.577654 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.680461 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.680515 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.680532 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.680557 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.680576 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.783299 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.783360 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.783376 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.783404 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.783423 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.830247 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.830357 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:23 crc kubenswrapper[4966]: E1217 08:22:23.830417 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.830252 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.830278 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:23 crc kubenswrapper[4966]: E1217 08:22:23.830611 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:23 crc kubenswrapper[4966]: E1217 08:22:23.830722 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:23 crc kubenswrapper[4966]: E1217 08:22:23.830785 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.886308 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.886355 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.886367 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.886385 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.886399 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.989962 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.990011 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.990026 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.990048 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:23 crc kubenswrapper[4966]: I1217 08:22:23.990063 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:23Z","lastTransitionTime":"2025-12-17T08:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.093249 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.093314 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.093337 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.093367 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.093391 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:24Z","lastTransitionTime":"2025-12-17T08:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.196358 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.196396 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.196406 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.196423 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.196435 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:24Z","lastTransitionTime":"2025-12-17T08:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.299226 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.299264 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.299276 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.299294 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.299306 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:24Z","lastTransitionTime":"2025-12-17T08:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.402111 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.402155 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.402170 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.402211 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.402236 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:24Z","lastTransitionTime":"2025-12-17T08:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.504096 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.504129 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.504137 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.504150 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.504158 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:24Z","lastTransitionTime":"2025-12-17T08:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.606806 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.606838 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.606846 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.606859 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.606880 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:24Z","lastTransitionTime":"2025-12-17T08:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.709231 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.709260 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.709268 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.709281 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.709289 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:24Z","lastTransitionTime":"2025-12-17T08:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.811357 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.811396 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.811407 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.811423 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.811435 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:24Z","lastTransitionTime":"2025-12-17T08:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.843722 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.861427 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.873745 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.885379 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.905157 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.913347 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.913397 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.913408 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.913427 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.913441 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:24Z","lastTransitionTime":"2025-12-17T08:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.919980 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.930233 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.945204 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.959510 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:22:11Z\\\",\\\"message\\\":\\\"2025-12-17T08:21:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1\\\\n2025-12-17T08:21:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1 to /host/opt/cni/bin/\\\\n2025-12-17T08:21:26Z [verbose] multus-daemon started\\\\n2025-12-17T08:21:26Z [verbose] Readiness Indicator file check\\\\n2025-12-17T08:22:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:22:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.973450 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:24 crc kubenswrapper[4966]: I1217 08:22:24.992660 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:57Z\\\",\\\"message\\\":\\\"_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1217 08:21:56.929996 6544 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1217 08:21:56.930024 6544 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1217 08:21:56.931578 6544 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:24Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.006242 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.016182 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.016240 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.016251 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.016271 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.016286 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.021385 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d221151-66d5-4d15-82d6-30df6d6e7ce7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.039945 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.053253 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.068634 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.084212 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.098756 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:25Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.118611 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.118648 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.118657 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.118672 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.118682 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.220997 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.221062 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.221072 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.221091 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.221100 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.323970 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.324035 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.324052 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.324075 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.324092 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.427297 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.427359 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.427379 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.427402 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.427419 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.529456 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.529492 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.529526 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.529544 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.529554 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.632493 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.632522 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.632533 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.632548 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.632559 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.735328 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.735396 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.735419 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.735449 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.735466 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.829695 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.829807 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.829695 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:25 crc kubenswrapper[4966]: E1217 08:22:25.829900 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:25 crc kubenswrapper[4966]: E1217 08:22:25.829801 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:25 crc kubenswrapper[4966]: E1217 08:22:25.829988 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.830198 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:25 crc kubenswrapper[4966]: E1217 08:22:25.830278 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.838121 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.838155 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.838169 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.838216 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.838228 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.940358 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.940402 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.940413 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.940429 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:25 crc kubenswrapper[4966]: I1217 08:22:25.940439 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:25Z","lastTransitionTime":"2025-12-17T08:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.042903 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.042928 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.042936 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.042949 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.042959 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.145394 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.145447 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.145463 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.145484 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.145499 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.248166 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.248205 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.248215 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.248230 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.248239 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.350664 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.350926 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.350994 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.351060 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.351114 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.453391 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.453433 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.453444 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.453461 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.453473 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.530942 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.531187 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:30.531150941 +0000 UTC m=+146.076220923 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.555515 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.555574 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.555591 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.555620 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.555637 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.632623 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.632838 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.632926 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.632970 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.632993 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.633007 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:23:30.632979203 +0000 UTC m=+146.178049185 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.632852 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.633063 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:23:30.633040846 +0000 UTC m=+146.178110818 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.633254 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.633335 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.633380 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.633431 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:23:30.633414817 +0000 UTC m=+146.178484799 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.633609 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.633632 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.633646 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:22:26 crc kubenswrapper[4966]: E1217 08:22:26.633726 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:23:30.633703475 +0000 UTC m=+146.178773417 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.659328 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.659401 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.659429 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.659460 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.659484 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.762754 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.762811 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.762828 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.762851 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.762867 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.830779 4966 scope.go:117] "RemoveContainer" containerID="2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.866077 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.866140 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.866162 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.866195 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.866217 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.969400 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.969446 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.969462 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.969484 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:26 crc kubenswrapper[4966]: I1217 08:22:26.969502 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:26Z","lastTransitionTime":"2025-12-17T08:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.073220 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.073362 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.073381 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.073404 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.073420 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.175981 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.176038 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.176055 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.176080 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.176100 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.279360 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.279432 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.279453 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.279480 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.279503 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.382213 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.382249 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.382287 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.382303 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.382312 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.415343 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.415409 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.415420 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.415434 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.415462 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.430463 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.434228 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.434331 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.434347 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.434369 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.434384 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.449473 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.457362 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.457431 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.457451 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.457477 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.457506 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.470982 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.474705 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.474748 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.474759 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.474776 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.474787 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.487910 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.491770 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.491967 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.491980 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.491998 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.492011 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.504695 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dbf2ad73-4183-43fc-86d4-935121ab90f9\\\",\\\"systemUUID\\\":\\\"34a07af6-4eb8-4386-b056-0160b9517aa1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:27Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.504850 4966 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.506930 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.506957 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.506968 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.506985 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.506997 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.608926 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.608970 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.608982 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.608999 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.609012 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.711449 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.711509 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.711526 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.711551 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.711570 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.813813 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.814129 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.814145 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.814165 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.814179 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.830260 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.830469 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.830806 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.830964 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.831308 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.831414 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.831722 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:27 crc kubenswrapper[4966]: E1217 08:22:27.831863 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.917343 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.917380 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.917395 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.917409 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:27 crc kubenswrapper[4966]: I1217 08:22:27.917418 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:27Z","lastTransitionTime":"2025-12-17T08:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.028724 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.028802 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.028817 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.028833 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.028841 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.135217 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.135259 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.135268 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.135283 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.135292 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.280390 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.280432 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.280443 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.280458 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.280470 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.295958 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/2.log" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.298725 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.299095 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.323956 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d221151-66d5-4d15-82d6-30df6d6e7ce7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.344853 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.382298 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.382326 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.382333 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.382350 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.382359 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.438482 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.465158 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.484147 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.484175 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.484185 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.484201 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.484211 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.504684 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.520922 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:57Z\\\",\\\"message\\\":\\\"_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1217 08:21:56.929996 6544 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1217 08:21:56.930024 6544 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1217 08:21:56.931578 6544 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.530069 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.540556 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.550511 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.561781 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.578717 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.586018 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.586041 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.586048 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.586061 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.586070 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.591370 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.601824 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.610460 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.621963 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.634950 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:22:11Z\\\",\\\"message\\\":\\\"2025-12-17T08:21:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1\\\\n2025-12-17T08:21:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1 to /host/opt/cni/bin/\\\\n2025-12-17T08:21:26Z [verbose] multus-daemon started\\\\n2025-12-17T08:21:26Z [verbose] Readiness Indicator file check\\\\n2025-12-17T08:22:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:22:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.644809 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.655951 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.689139 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.689184 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.689196 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.689213 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.689224 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.791180 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.791219 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.791230 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.791247 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.791258 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.893486 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.893543 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.893555 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.893578 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.893595 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.996259 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.996289 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.996298 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.996310 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:28 crc kubenswrapper[4966]: I1217 08:22:28.996319 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:28Z","lastTransitionTime":"2025-12-17T08:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.098601 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.098653 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.098663 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.098677 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.098687 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:29Z","lastTransitionTime":"2025-12-17T08:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.200706 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.200977 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.201054 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.201148 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.201248 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:29Z","lastTransitionTime":"2025-12-17T08:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.303416 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.303454 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.303465 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.303482 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.303494 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:29Z","lastTransitionTime":"2025-12-17T08:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.406113 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.406159 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.406174 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.406194 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.406208 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:29Z","lastTransitionTime":"2025-12-17T08:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.508479 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.508865 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.509094 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.509294 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.509470 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:29Z","lastTransitionTime":"2025-12-17T08:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.612675 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.612735 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.612752 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.612784 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.612801 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:29Z","lastTransitionTime":"2025-12-17T08:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.716012 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.716070 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.716095 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.716127 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.716149 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:29Z","lastTransitionTime":"2025-12-17T08:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.818810 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.818845 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.818856 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.818898 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.818910 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:29Z","lastTransitionTime":"2025-12-17T08:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.829705 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:29 crc kubenswrapper[4966]: E1217 08:22:29.829811 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.829979 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:29 crc kubenswrapper[4966]: E1217 08:22:29.830024 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.830132 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:29 crc kubenswrapper[4966]: E1217 08:22:29.830181 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.830173 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:29 crc kubenswrapper[4966]: E1217 08:22:29.830230 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.928058 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.928112 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.928128 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.928152 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:29 crc kubenswrapper[4966]: I1217 08:22:29.928169 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:29Z","lastTransitionTime":"2025-12-17T08:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.031128 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.031197 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.031220 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.031250 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.031271 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.134040 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.134069 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.134081 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.134105 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.134116 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.237604 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.237648 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.237657 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.237673 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.237685 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.307409 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/3.log" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.308158 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/2.log" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.312297 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" exitCode=1 Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.312353 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.312433 4966 scope.go:117] "RemoveContainer" containerID="2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.314776 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:22:30 crc kubenswrapper[4966]: E1217 08:22:30.321097 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.334728 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dp4dv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ca416e0-2aab-40ba-b42f-a8ed3afbbafb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef3fb2ab895f757884dea0cc410f7f7ee9cca8204f6684b02d2f6583a45847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xsjb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dp4dv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.343015 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.343056 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.343071 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.343090 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.343103 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.351285 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f703caec-d8d0-4e72-b58a-987e69356984\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff0cc71876b011ddd9079a233174f57928ae8b1cee4da0d1beee03261b2d7b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lx2dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.369628 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqk4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac204c30-92c9-4b1d-9cda-bef16b7a0008\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:22:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:22:11Z\\\",\\\"message\\\":\\\"2025-12-17T08:21:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1\\\\n2025-12-17T08:21:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_93a27329-68da-48c5-85ec-d3d615544bd1 to /host/opt/cni/bin/\\\\n2025-12-17T08:21:26Z [verbose] multus-daemon started\\\\n2025-12-17T08:21:26Z [verbose] Readiness Indicator file check\\\\n2025-12-17T08:22:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:22:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24lmz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqk4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.381964 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b447x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42cd7025-811b-4b83-b1df-d43919ce03de\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8cb75d27b5b48e480cc0cca7dec91280b9813c30e7f6ff830d771f66b68a7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vttjv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:25Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b447x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.393657 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"216f8233-5cc0-4733-b27e-69f9312d9a94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b7ee6fece8cd2fa1898a54cf1fad5378c1d064a0b5b86bfb2e79ec2dcc5edf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1820e0896faae209320c5412255d7bedfd530d120538de0b21895a9d74a9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2n4cv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mw9pk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.413579 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00ec6e41-e778-4f9e-98b8-593f1ec04a03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"le observer\\\\nW1217 08:21:21.980172 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1217 08:21:21.980302 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1217 08:21:21.981244 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3252013971/tls.crt::/tmp/serving-cert-3252013971/tls.key\\\\\\\"\\\\nI1217 08:21:22.428773 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1217 08:21:22.430936 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1217 08:21:22.430952 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1217 08:21:22.430973 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1217 08:21:22.430978 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1217 08:21:22.436286 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1217 08:21:22.436609 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1217 08:21:22.436688 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1217 08:21:22.436744 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1217 08:21:22.436765 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1217 08:21:22.436784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1217 08:21:22.436380 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1217 08:21:22.438459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.425449 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.443487 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b7d5ef4-9acc-4b53-b830-2e5a5c906864\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf1bafb7c661237ba5efccb9bbd3f13b2745701846c48a18ee291998da77549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2a883297549f77969779707c9610c8f68322f78b99b09255adbcdc7adcfc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f394743f900fb31199fc19dd42feab3b325662dde14260b9ebb0c239345fac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://931f573d0cdbf482d3d4cafd0510fb56af64157d88192e93fcf6078c897538f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd117f30d3cb58f4448fa3f53a28b3d6e65d63a7adaa17ce338738d91881f0b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://683fcd7150e451565125953dc2a75457981a928f8482ff8feded93f036214cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfba49903d49cfac51db41a24dc686dfc3327c54bca98e36d14f7d207fb1eeb4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5d192b02fd68f39c16e33fe842b409bfbb67059d8954cd50578cce5ecaef77b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.451944 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.451980 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.451992 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.452009 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.452021 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.469146 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.482190 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2a70a577e5885f2ab264eb6e46babad9840bb49d75048718335e7899043cf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.495678 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.512794 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"865dfc88-97a7-4cdb-9564-f46cfadae5dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2996498acc63741b4e18bac0fefc63efa52b848220f532c7b78dfa356a79a461\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:21:57Z\\\",\\\"message\\\":\\\"_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1217 08:21:56.929996 6544 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1217 08:21:56.930024 6544 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1217 08:21:56.931578 6544 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:21:56Z is after 2025-08-24T17:21:41Z]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-17T08:22:29Z\\\",\\\"message\\\":\\\"8] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1217 08:22:29.002649 6978 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1217 08:22:29.002658 6978 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1217 08:22:29.002683 6978 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1217 08:22:29.002719 6978 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:28Z is after 2025-08-24T17:21:41Z]\\\\nI1217 08:22:29.002593 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-17T08:22:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnvtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6lq7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.526118 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nqmws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cf617b8-bd09-49de-b188-9c86d5fafd57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdrrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nqmws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.539052 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d221151-66d5-4d15-82d6-30df6d6e7ce7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24024c52832dc8d0c8e421fc0ebbefaf8ac74613ff84814a0033db1048b8c404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6037400aaed90b823a1b85a3153bad9b85c7c1335780bcda33905715cff5f616\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fb55ef486cc09304b27239b50d1e1a46567aae1c8a71a6de9e95274034e91df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70a5221cf650c44fe2467480bc86cdf323df2cd18fa556412147aa0ed9477b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.551062 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1390a059-4556-4def-8615-2a1f6d116455\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1a6912dbd2ca84c97d0e42acd620455f9698992ddd2090b7257fe99a8912eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c96da79bd7f751a863cc29dffcf5fd437ef3024c0a975eeb7bf244db5176208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de9badeabff83bb01b051c3e5aa9266c5384be380f36676624518707cab5dc0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.554004 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.554028 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.554036 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.554070 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.554080 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.563510 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfa367197b100787c8ee58510314bb1263c6c7e738af92c80c17129fd25f7d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.573446 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1318d4ef9600080d7a5ed3f4d24ed02a3588666834898ceb4941900d06363628\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b21e5649f6550c1c07b7870d6fe3fa7ef74aac4806ba21883ca1032a510aca1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.590514 4966 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ab08e5a-5118-46ba-9b61-d0e5939c3083\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-17T08:21:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c434a6d419af4c321469af3e898ea125085e9dba07f55c45424b734f91a849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-17T08:21:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4fde0cd78c550ebbd1ae0dfeb2ddb4ee05d292a56f82dc312f64cfd1af9699\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb025e6eab7e0fb50b4c3d473f548a1607c349ed1ad6cba72c0e9a4649a6ca7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fc926de9a32e15471d86b0f7891d4a98e1bd6faaa73e225ae5602cc4c4319fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://effb53c03e399fdd75c8b39a7e02d5a6bf0f58da44aebefbad44ec1d5ca23acc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58068ca9de51298556fbc55d880ebdf9e4231173e8e5589370cff8b201a4328e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcb42068591d212f2ffb84693ff2f5ba8ab40203903ec7310f2abebfa4129d09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-17T08:21:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-17T08:21:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9k9d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-17T08:21:23Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4q4sn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-17T08:22:30Z is after 2025-08-24T17:21:41Z" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.656518 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.656557 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.656585 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.656602 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.656612 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.758948 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.759020 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.759089 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.759122 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.759144 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.861448 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.861534 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.861551 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.861576 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.861592 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.964379 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.964455 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.964473 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.964489 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:30 crc kubenswrapper[4966]: I1217 08:22:30.964499 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:30Z","lastTransitionTime":"2025-12-17T08:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.067922 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.067986 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.068004 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.068031 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.068051 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.171149 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.171200 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.171219 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.171243 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.171261 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.274617 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.274653 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.274663 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.274676 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.274685 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.317415 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/3.log" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.377544 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.377585 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.377595 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.377611 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.377623 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.480298 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.480362 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.480376 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.480394 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.480406 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.583126 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.583169 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.583182 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.583200 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.583214 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.686619 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.686693 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.686728 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.686762 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.686785 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.789098 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.789137 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.789148 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.789167 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.789179 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.829629 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.829676 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.829730 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:31 crc kubenswrapper[4966]: E1217 08:22:31.829888 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.829986 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:31 crc kubenswrapper[4966]: E1217 08:22:31.830073 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:31 crc kubenswrapper[4966]: E1217 08:22:31.830138 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:31 crc kubenswrapper[4966]: E1217 08:22:31.830338 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.891433 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.891459 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.891467 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.891480 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.891488 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.993963 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.994028 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.994040 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.994057 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:31 crc kubenswrapper[4966]: I1217 08:22:31.994069 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:31Z","lastTransitionTime":"2025-12-17T08:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.097115 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.097196 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.097223 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.097256 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.097282 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:32Z","lastTransitionTime":"2025-12-17T08:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.200005 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.200073 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.200121 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.200150 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.200168 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:32Z","lastTransitionTime":"2025-12-17T08:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.303270 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.303315 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.303326 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.303344 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.303356 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:32Z","lastTransitionTime":"2025-12-17T08:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.406650 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.406695 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.406706 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.406722 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.406732 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:32Z","lastTransitionTime":"2025-12-17T08:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.510133 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.510202 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.510226 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.510255 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.510279 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:32Z","lastTransitionTime":"2025-12-17T08:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.612476 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.612531 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.612542 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.612556 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.612566 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:32Z","lastTransitionTime":"2025-12-17T08:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.715142 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.715174 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.715186 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.715203 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.715216 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:32Z","lastTransitionTime":"2025-12-17T08:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.818038 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.818090 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.818103 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.818123 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.818135 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:32Z","lastTransitionTime":"2025-12-17T08:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.920723 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.920779 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.920796 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.920827 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:32 crc kubenswrapper[4966]: I1217 08:22:32.920848 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:32Z","lastTransitionTime":"2025-12-17T08:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.023193 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.023229 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.023239 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.023253 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.023263 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.126616 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.126672 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.126682 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.126702 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.126717 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.230731 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.230792 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.230801 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.230819 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.230829 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.333226 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.333262 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.333270 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.333284 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.333295 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.436712 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.436773 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.436789 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.436814 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.436833 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.540237 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.540293 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.540306 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.540325 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.540337 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.642479 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.642526 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.642537 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.642555 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.642570 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.745724 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.745791 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.745812 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.745842 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.745863 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.829546 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.829653 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.829556 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:33 crc kubenswrapper[4966]: E1217 08:22:33.829706 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.829653 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:33 crc kubenswrapper[4966]: E1217 08:22:33.829794 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:33 crc kubenswrapper[4966]: E1217 08:22:33.829953 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:33 crc kubenswrapper[4966]: E1217 08:22:33.829987 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.848450 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.848494 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.848507 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.848524 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.848538 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.950431 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.950485 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.950495 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.950513 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:33 crc kubenswrapper[4966]: I1217 08:22:33.950526 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:33Z","lastTransitionTime":"2025-12-17T08:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.052550 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.052588 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.052597 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.052611 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.052622 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.154961 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.154991 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.155001 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.155041 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.155055 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.257749 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.257801 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.257818 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.257840 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.257857 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.360231 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.360271 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.360280 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.360294 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.360304 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.464166 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.464274 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.464290 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.464312 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.464351 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.568190 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.568229 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.568237 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.568252 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.568261 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.671126 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.671173 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.671185 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.671203 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.671213 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.774166 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.774208 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.774226 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.774242 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.774254 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.877112 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.877189 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.877202 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.877219 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.877231 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.911859 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=38.911838289 podStartE2EDuration="38.911838289s" podCreationTimestamp="2025-12-17 08:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:34.910802869 +0000 UTC m=+90.455872811" watchObservedRunningTime="2025-12-17 08:22:34.911838289 +0000 UTC m=+90.456908231" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.936627 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=72.936611709 podStartE2EDuration="1m12.936611709s" podCreationTimestamp="2025-12-17 08:21:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:34.936229779 +0000 UTC m=+90.481299741" watchObservedRunningTime="2025-12-17 08:22:34.936611709 +0000 UTC m=+90.481681651" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.979268 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.979299 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.979309 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.979325 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.979314 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=72.979303179 podStartE2EDuration="1m12.979303179s" podCreationTimestamp="2025-12-17 08:21:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:34.979150263 +0000 UTC m=+90.524220205" watchObservedRunningTime="2025-12-17 08:22:34.979303179 +0000 UTC m=+90.524373121" Dec 17 08:22:34 crc kubenswrapper[4966]: I1217 08:22:34.979337 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:34Z","lastTransitionTime":"2025-12-17T08:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.028901 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-4q4sn" podStartSLOduration=72.028854119 podStartE2EDuration="1m12.028854119s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:35.028679604 +0000 UTC m=+90.573749566" watchObservedRunningTime="2025-12-17 08:22:35.028854119 +0000 UTC m=+90.573924061" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.039850 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-b447x" podStartSLOduration=72.039832303 podStartE2EDuration="1m12.039832303s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:35.039234265 +0000 UTC m=+90.584304217" watchObservedRunningTime="2025-12-17 08:22:35.039832303 +0000 UTC m=+90.584902255" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.065523 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=73.06550445 podStartE2EDuration="1m13.06550445s" podCreationTimestamp="2025-12-17 08:21:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:35.064535991 +0000 UTC m=+90.609605943" watchObservedRunningTime="2025-12-17 08:22:35.06550445 +0000 UTC m=+90.610574392" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.065700 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mw9pk" podStartSLOduration=72.065676385 podStartE2EDuration="1m12.065676385s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:35.049477468 +0000 UTC m=+90.594547420" watchObservedRunningTime="2025-12-17 08:22:35.065676385 +0000 UTC m=+90.610746327" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.081862 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.081936 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.081949 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.081968 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.081989 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:35Z","lastTransitionTime":"2025-12-17T08:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.100553 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-dp4dv" podStartSLOduration=72.100533763 podStartE2EDuration="1m12.100533763s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:35.087741526 +0000 UTC m=+90.632811468" watchObservedRunningTime="2025-12-17 08:22:35.100533763 +0000 UTC m=+90.645603705" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.116911 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podStartSLOduration=72.116891465 podStartE2EDuration="1m12.116891465s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:35.101629505 +0000 UTC m=+90.646699477" watchObservedRunningTime="2025-12-17 08:22:35.116891465 +0000 UTC m=+90.661961417" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.184620 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.184650 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.184658 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.184671 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.184680 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:35Z","lastTransitionTime":"2025-12-17T08:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.287346 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.287392 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.287403 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.287420 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.287433 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:35Z","lastTransitionTime":"2025-12-17T08:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.389702 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.389762 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.389774 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.389792 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.389803 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:35Z","lastTransitionTime":"2025-12-17T08:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.492296 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.492340 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.492351 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.492367 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.492379 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:35Z","lastTransitionTime":"2025-12-17T08:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.594425 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.594469 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.594481 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.594498 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.594511 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:35Z","lastTransitionTime":"2025-12-17T08:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.697649 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.697699 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.697713 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.697735 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.697750 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:35Z","lastTransitionTime":"2025-12-17T08:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.800051 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.800100 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.800116 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.800136 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.800148 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:35Z","lastTransitionTime":"2025-12-17T08:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.830009 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.830062 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.830062 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.830062 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:35 crc kubenswrapper[4966]: E1217 08:22:35.830176 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:35 crc kubenswrapper[4966]: E1217 08:22:35.830454 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:35 crc kubenswrapper[4966]: E1217 08:22:35.830553 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:35 crc kubenswrapper[4966]: E1217 08:22:35.830612 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.902840 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.902923 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.902940 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.902964 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:35 crc kubenswrapper[4966]: I1217 08:22:35.902982 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:35Z","lastTransitionTime":"2025-12-17T08:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.005497 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.005547 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.005558 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.005576 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.005587 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.108347 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.108404 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.108426 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.108456 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.108476 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.212116 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.212194 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.212211 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.212236 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.212250 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.314136 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.314174 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.314182 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.314198 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.314208 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.416587 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.416623 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.416634 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.416650 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.416661 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.519366 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.519454 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.519471 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.519491 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.519502 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.621622 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.621660 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.621670 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.621687 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.621699 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.723954 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.724022 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.724035 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.724059 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.724076 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.826985 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.827050 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.827063 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.827081 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.827093 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.846779 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-lqk4j" podStartSLOduration=73.846754758 podStartE2EDuration="1m13.846754758s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:35.118258616 +0000 UTC m=+90.663328578" watchObservedRunningTime="2025-12-17 08:22:36.846754758 +0000 UTC m=+92.391824760" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.847839 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.929374 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.929428 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.929439 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.929456 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:36 crc kubenswrapper[4966]: I1217 08:22:36.929487 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:36Z","lastTransitionTime":"2025-12-17T08:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.031462 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.031498 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.031528 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.031546 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.031558 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:37Z","lastTransitionTime":"2025-12-17T08:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.133809 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.133841 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.133853 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.133882 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.133894 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:37Z","lastTransitionTime":"2025-12-17T08:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.235766 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.235800 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.235812 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.235828 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.235840 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:37Z","lastTransitionTime":"2025-12-17T08:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.338776 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.338831 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.338852 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.338916 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.338943 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:37Z","lastTransitionTime":"2025-12-17T08:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.440936 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.441015 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.441033 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.441055 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.441074 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:37Z","lastTransitionTime":"2025-12-17T08:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.543080 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.543124 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.543138 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.543161 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.543179 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:37Z","lastTransitionTime":"2025-12-17T08:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.644864 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.644949 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.644962 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.644982 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.644994 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:37Z","lastTransitionTime":"2025-12-17T08:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.748497 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.748549 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.748559 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.748576 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.748587 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:37Z","lastTransitionTime":"2025-12-17T08:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.818858 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.818951 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.818965 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.818989 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.819004 4966 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-17T08:22:37Z","lastTransitionTime":"2025-12-17T08:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.829993 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.830017 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.829993 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:37 crc kubenswrapper[4966]: E1217 08:22:37.830251 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:37 crc kubenswrapper[4966]: E1217 08:22:37.830146 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.830316 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:37 crc kubenswrapper[4966]: E1217 08:22:37.830381 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:37 crc kubenswrapper[4966]: E1217 08:22:37.830448 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.864088 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j"] Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.864791 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.867462 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.867609 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.868254 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.870414 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.882420 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.882399154 podStartE2EDuration="1.882399154s" podCreationTimestamp="2025-12-17 08:22:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:37.88226906 +0000 UTC m=+93.427339012" watchObservedRunningTime="2025-12-17 08:22:37.882399154 +0000 UTC m=+93.427469096" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.964184 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/762c8f9d-53de-4510-b634-cc5ccba50f50-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.964230 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/762c8f9d-53de-4510-b634-cc5ccba50f50-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.964257 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/762c8f9d-53de-4510-b634-cc5ccba50f50-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.964310 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/762c8f9d-53de-4510-b634-cc5ccba50f50-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:37 crc kubenswrapper[4966]: I1217 08:22:37.964334 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/762c8f9d-53de-4510-b634-cc5ccba50f50-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.065534 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/762c8f9d-53de-4510-b634-cc5ccba50f50-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.065580 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/762c8f9d-53de-4510-b634-cc5ccba50f50-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.065604 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/762c8f9d-53de-4510-b634-cc5ccba50f50-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.065655 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/762c8f9d-53de-4510-b634-cc5ccba50f50-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.065681 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/762c8f9d-53de-4510-b634-cc5ccba50f50-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.065833 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/762c8f9d-53de-4510-b634-cc5ccba50f50-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.066147 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/762c8f9d-53de-4510-b634-cc5ccba50f50-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.066897 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/762c8f9d-53de-4510-b634-cc5ccba50f50-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.072092 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/762c8f9d-53de-4510-b634-cc5ccba50f50-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.088688 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/762c8f9d-53de-4510-b634-cc5ccba50f50-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wp28j\" (UID: \"762c8f9d-53de-4510-b634-cc5ccba50f50\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.182978 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.346397 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" event={"ID":"762c8f9d-53de-4510-b634-cc5ccba50f50","Type":"ContainerStarted","Data":"d3e86d27d6ae2f8201e0aa995d0a12e9298ad0ca3852533b813f85aea644586d"} Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.346487 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" event={"ID":"762c8f9d-53de-4510-b634-cc5ccba50f50","Type":"ContainerStarted","Data":"d9a8d555ffe28fd860592070dd2476cbd8360a9e994a303f7f92a1ea92dfea8e"} Dec 17 08:22:38 crc kubenswrapper[4966]: I1217 08:22:38.362669 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wp28j" podStartSLOduration=75.362641464 podStartE2EDuration="1m15.362641464s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:22:38.362544841 +0000 UTC m=+93.907614783" watchObservedRunningTime="2025-12-17 08:22:38.362641464 +0000 UTC m=+93.907711406" Dec 17 08:22:39 crc kubenswrapper[4966]: I1217 08:22:39.830326 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:39 crc kubenswrapper[4966]: I1217 08:22:39.830433 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:39 crc kubenswrapper[4966]: E1217 08:22:39.830479 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:39 crc kubenswrapper[4966]: E1217 08:22:39.830567 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:39 crc kubenswrapper[4966]: I1217 08:22:39.830350 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:39 crc kubenswrapper[4966]: E1217 08:22:39.830640 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:39 crc kubenswrapper[4966]: I1217 08:22:39.830707 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:39 crc kubenswrapper[4966]: E1217 08:22:39.830771 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:41 crc kubenswrapper[4966]: I1217 08:22:41.603617 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:41 crc kubenswrapper[4966]: E1217 08:22:41.603817 4966 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:22:41 crc kubenswrapper[4966]: E1217 08:22:41.603943 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs podName:6cf617b8-bd09-49de-b188-9c86d5fafd57 nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.603919101 +0000 UTC m=+161.148989093 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs") pod "network-metrics-daemon-nqmws" (UID: "6cf617b8-bd09-49de-b188-9c86d5fafd57") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 17 08:22:41 crc kubenswrapper[4966]: I1217 08:22:41.830230 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:41 crc kubenswrapper[4966]: I1217 08:22:41.830271 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:41 crc kubenswrapper[4966]: I1217 08:22:41.830335 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:41 crc kubenswrapper[4966]: E1217 08:22:41.830377 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:41 crc kubenswrapper[4966]: E1217 08:22:41.830460 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:41 crc kubenswrapper[4966]: E1217 08:22:41.830520 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:41 crc kubenswrapper[4966]: I1217 08:22:41.830261 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:41 crc kubenswrapper[4966]: E1217 08:22:41.831107 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:42 crc kubenswrapper[4966]: I1217 08:22:42.830836 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:22:42 crc kubenswrapper[4966]: E1217 08:22:42.831122 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" Dec 17 08:22:43 crc kubenswrapper[4966]: I1217 08:22:43.830076 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:43 crc kubenswrapper[4966]: I1217 08:22:43.830356 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:43 crc kubenswrapper[4966]: I1217 08:22:43.830375 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:43 crc kubenswrapper[4966]: I1217 08:22:43.830173 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:43 crc kubenswrapper[4966]: E1217 08:22:43.830682 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:43 crc kubenswrapper[4966]: E1217 08:22:43.830835 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:43 crc kubenswrapper[4966]: E1217 08:22:43.831313 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:43 crc kubenswrapper[4966]: E1217 08:22:43.831897 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:45 crc kubenswrapper[4966]: I1217 08:22:45.829837 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:45 crc kubenswrapper[4966]: E1217 08:22:45.830734 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:45 crc kubenswrapper[4966]: I1217 08:22:45.830055 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:45 crc kubenswrapper[4966]: E1217 08:22:45.831329 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:45 crc kubenswrapper[4966]: I1217 08:22:45.830018 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:45 crc kubenswrapper[4966]: E1217 08:22:45.831614 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:45 crc kubenswrapper[4966]: I1217 08:22:45.830155 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:45 crc kubenswrapper[4966]: E1217 08:22:45.831899 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:47 crc kubenswrapper[4966]: I1217 08:22:47.830173 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:47 crc kubenswrapper[4966]: I1217 08:22:47.830235 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:47 crc kubenswrapper[4966]: I1217 08:22:47.830271 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:47 crc kubenswrapper[4966]: I1217 08:22:47.830491 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:47 crc kubenswrapper[4966]: E1217 08:22:47.831171 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:47 crc kubenswrapper[4966]: E1217 08:22:47.831317 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:47 crc kubenswrapper[4966]: E1217 08:22:47.831412 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:47 crc kubenswrapper[4966]: E1217 08:22:47.831447 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:49 crc kubenswrapper[4966]: I1217 08:22:49.829703 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:49 crc kubenswrapper[4966]: I1217 08:22:49.829733 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:49 crc kubenswrapper[4966]: I1217 08:22:49.830541 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:49 crc kubenswrapper[4966]: I1217 08:22:49.830569 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:49 crc kubenswrapper[4966]: E1217 08:22:49.830681 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:49 crc kubenswrapper[4966]: E1217 08:22:49.830743 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:49 crc kubenswrapper[4966]: E1217 08:22:49.830818 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:49 crc kubenswrapper[4966]: E1217 08:22:49.830895 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:51 crc kubenswrapper[4966]: I1217 08:22:51.830104 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:51 crc kubenswrapper[4966]: I1217 08:22:51.830183 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:51 crc kubenswrapper[4966]: I1217 08:22:51.830234 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:51 crc kubenswrapper[4966]: I1217 08:22:51.830331 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:51 crc kubenswrapper[4966]: E1217 08:22:51.831656 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:51 crc kubenswrapper[4966]: E1217 08:22:51.832386 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:51 crc kubenswrapper[4966]: E1217 08:22:51.832960 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:51 crc kubenswrapper[4966]: E1217 08:22:51.833361 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:53 crc kubenswrapper[4966]: I1217 08:22:53.830212 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:53 crc kubenswrapper[4966]: I1217 08:22:53.830248 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:53 crc kubenswrapper[4966]: I1217 08:22:53.830349 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:53 crc kubenswrapper[4966]: I1217 08:22:53.830460 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:53 crc kubenswrapper[4966]: E1217 08:22:53.831913 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:53 crc kubenswrapper[4966]: E1217 08:22:53.831707 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:53 crc kubenswrapper[4966]: E1217 08:22:53.831464 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:53 crc kubenswrapper[4966]: E1217 08:22:53.832056 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:54 crc kubenswrapper[4966]: I1217 08:22:54.832333 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:22:54 crc kubenswrapper[4966]: E1217 08:22:54.832521 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" Dec 17 08:22:55 crc kubenswrapper[4966]: I1217 08:22:55.830133 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:55 crc kubenswrapper[4966]: I1217 08:22:55.830191 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:55 crc kubenswrapper[4966]: I1217 08:22:55.830226 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:55 crc kubenswrapper[4966]: I1217 08:22:55.830226 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:55 crc kubenswrapper[4966]: E1217 08:22:55.830859 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:55 crc kubenswrapper[4966]: E1217 08:22:55.830736 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:55 crc kubenswrapper[4966]: E1217 08:22:55.830938 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:55 crc kubenswrapper[4966]: E1217 08:22:55.830634 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:57 crc kubenswrapper[4966]: I1217 08:22:57.829546 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:57 crc kubenswrapper[4966]: I1217 08:22:57.829667 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:57 crc kubenswrapper[4966]: E1217 08:22:57.829733 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:57 crc kubenswrapper[4966]: I1217 08:22:57.829772 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:57 crc kubenswrapper[4966]: E1217 08:22:57.829958 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:22:57 crc kubenswrapper[4966]: E1217 08:22:57.830021 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:57 crc kubenswrapper[4966]: I1217 08:22:57.829608 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:57 crc kubenswrapper[4966]: E1217 08:22:57.831281 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:58 crc kubenswrapper[4966]: I1217 08:22:58.409901 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/1.log" Dec 17 08:22:58 crc kubenswrapper[4966]: I1217 08:22:58.410352 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/0.log" Dec 17 08:22:58 crc kubenswrapper[4966]: I1217 08:22:58.410415 4966 generic.go:334] "Generic (PLEG): container finished" podID="ac204c30-92c9-4b1d-9cda-bef16b7a0008" containerID="2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1" exitCode=1 Dec 17 08:22:58 crc kubenswrapper[4966]: I1217 08:22:58.410455 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqk4j" event={"ID":"ac204c30-92c9-4b1d-9cda-bef16b7a0008","Type":"ContainerDied","Data":"2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1"} Dec 17 08:22:58 crc kubenswrapper[4966]: I1217 08:22:58.410501 4966 scope.go:117] "RemoveContainer" containerID="6d13bf10704123cdc62921c94c2c3f4e6be7440ae2ee40d3694b548a19aa7e74" Dec 17 08:22:58 crc kubenswrapper[4966]: I1217 08:22:58.411182 4966 scope.go:117] "RemoveContainer" containerID="2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1" Dec 17 08:22:58 crc kubenswrapper[4966]: E1217 08:22:58.411469 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-lqk4j_openshift-multus(ac204c30-92c9-4b1d-9cda-bef16b7a0008)\"" pod="openshift-multus/multus-lqk4j" podUID="ac204c30-92c9-4b1d-9cda-bef16b7a0008" Dec 17 08:22:59 crc kubenswrapper[4966]: I1217 08:22:59.416318 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/1.log" Dec 17 08:22:59 crc kubenswrapper[4966]: I1217 08:22:59.829925 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:22:59 crc kubenswrapper[4966]: I1217 08:22:59.829925 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:22:59 crc kubenswrapper[4966]: I1217 08:22:59.830073 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:22:59 crc kubenswrapper[4966]: E1217 08:22:59.830207 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:22:59 crc kubenswrapper[4966]: E1217 08:22:59.830498 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:22:59 crc kubenswrapper[4966]: E1217 08:22:59.830584 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:22:59 crc kubenswrapper[4966]: I1217 08:22:59.830776 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:22:59 crc kubenswrapper[4966]: E1217 08:22:59.831115 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:01 crc kubenswrapper[4966]: I1217 08:23:01.830442 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:01 crc kubenswrapper[4966]: I1217 08:23:01.830535 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:01 crc kubenswrapper[4966]: I1217 08:23:01.830562 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:01 crc kubenswrapper[4966]: E1217 08:23:01.830725 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:01 crc kubenswrapper[4966]: I1217 08:23:01.830742 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:01 crc kubenswrapper[4966]: E1217 08:23:01.830849 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:01 crc kubenswrapper[4966]: E1217 08:23:01.830964 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:01 crc kubenswrapper[4966]: E1217 08:23:01.831025 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:03 crc kubenswrapper[4966]: I1217 08:23:03.829669 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:03 crc kubenswrapper[4966]: I1217 08:23:03.829731 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:03 crc kubenswrapper[4966]: I1217 08:23:03.829780 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:03 crc kubenswrapper[4966]: I1217 08:23:03.829692 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:03 crc kubenswrapper[4966]: E1217 08:23:03.829909 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:03 crc kubenswrapper[4966]: E1217 08:23:03.829985 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:03 crc kubenswrapper[4966]: E1217 08:23:03.830077 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:03 crc kubenswrapper[4966]: E1217 08:23:03.830150 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:04 crc kubenswrapper[4966]: E1217 08:23:04.726251 4966 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 17 08:23:04 crc kubenswrapper[4966]: E1217 08:23:04.899083 4966 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 17 08:23:05 crc kubenswrapper[4966]: I1217 08:23:05.830383 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:05 crc kubenswrapper[4966]: I1217 08:23:05.830404 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:05 crc kubenswrapper[4966]: I1217 08:23:05.830435 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:05 crc kubenswrapper[4966]: E1217 08:23:05.830512 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:05 crc kubenswrapper[4966]: I1217 08:23:05.830385 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:05 crc kubenswrapper[4966]: E1217 08:23:05.830601 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:05 crc kubenswrapper[4966]: E1217 08:23:05.830650 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:05 crc kubenswrapper[4966]: E1217 08:23:05.830709 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:07 crc kubenswrapper[4966]: I1217 08:23:07.830202 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:07 crc kubenswrapper[4966]: I1217 08:23:07.830278 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:07 crc kubenswrapper[4966]: I1217 08:23:07.830211 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:07 crc kubenswrapper[4966]: I1217 08:23:07.830327 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:07 crc kubenswrapper[4966]: E1217 08:23:07.830359 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:07 crc kubenswrapper[4966]: E1217 08:23:07.830434 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:07 crc kubenswrapper[4966]: E1217 08:23:07.830487 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:07 crc kubenswrapper[4966]: E1217 08:23:07.830519 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:08 crc kubenswrapper[4966]: I1217 08:23:08.831216 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:23:08 crc kubenswrapper[4966]: E1217 08:23:08.831503 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6lq7n_openshift-ovn-kubernetes(865dfc88-97a7-4cdb-9564-f46cfadae5dd)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" Dec 17 08:23:09 crc kubenswrapper[4966]: I1217 08:23:09.829457 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:09 crc kubenswrapper[4966]: I1217 08:23:09.829535 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:09 crc kubenswrapper[4966]: I1217 08:23:09.829455 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:09 crc kubenswrapper[4966]: E1217 08:23:09.829597 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:09 crc kubenswrapper[4966]: I1217 08:23:09.829467 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:09 crc kubenswrapper[4966]: E1217 08:23:09.829670 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:09 crc kubenswrapper[4966]: E1217 08:23:09.829805 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:09 crc kubenswrapper[4966]: E1217 08:23:09.829902 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:09 crc kubenswrapper[4966]: E1217 08:23:09.901161 4966 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 17 08:23:11 crc kubenswrapper[4966]: I1217 08:23:11.830114 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:11 crc kubenswrapper[4966]: I1217 08:23:11.830189 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:11 crc kubenswrapper[4966]: I1217 08:23:11.830199 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:11 crc kubenswrapper[4966]: I1217 08:23:11.830121 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:11 crc kubenswrapper[4966]: E1217 08:23:11.830285 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:11 crc kubenswrapper[4966]: E1217 08:23:11.830427 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:11 crc kubenswrapper[4966]: E1217 08:23:11.830525 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:11 crc kubenswrapper[4966]: E1217 08:23:11.830614 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:12 crc kubenswrapper[4966]: I1217 08:23:12.829818 4966 scope.go:117] "RemoveContainer" containerID="2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1" Dec 17 08:23:13 crc kubenswrapper[4966]: I1217 08:23:13.461272 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/1.log" Dec 17 08:23:13 crc kubenswrapper[4966]: I1217 08:23:13.461600 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqk4j" event={"ID":"ac204c30-92c9-4b1d-9cda-bef16b7a0008","Type":"ContainerStarted","Data":"66f24290075e5dc241bbf1d65160b46bebcbca3bbfb1204419f82204f5c2dde4"} Dec 17 08:23:13 crc kubenswrapper[4966]: I1217 08:23:13.830277 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:13 crc kubenswrapper[4966]: I1217 08:23:13.830332 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:13 crc kubenswrapper[4966]: I1217 08:23:13.830290 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:13 crc kubenswrapper[4966]: E1217 08:23:13.830405 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:13 crc kubenswrapper[4966]: E1217 08:23:13.830532 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:13 crc kubenswrapper[4966]: I1217 08:23:13.830549 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:13 crc kubenswrapper[4966]: E1217 08:23:13.830591 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:13 crc kubenswrapper[4966]: E1217 08:23:13.830651 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:14 crc kubenswrapper[4966]: E1217 08:23:14.901713 4966 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 17 08:23:15 crc kubenswrapper[4966]: I1217 08:23:15.830003 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:15 crc kubenswrapper[4966]: I1217 08:23:15.830041 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:15 crc kubenswrapper[4966]: I1217 08:23:15.830075 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:15 crc kubenswrapper[4966]: I1217 08:23:15.830021 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:15 crc kubenswrapper[4966]: E1217 08:23:15.830141 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:15 crc kubenswrapper[4966]: E1217 08:23:15.830203 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:15 crc kubenswrapper[4966]: E1217 08:23:15.830364 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:15 crc kubenswrapper[4966]: E1217 08:23:15.830502 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:17 crc kubenswrapper[4966]: I1217 08:23:17.830143 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:17 crc kubenswrapper[4966]: I1217 08:23:17.830229 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:17 crc kubenswrapper[4966]: I1217 08:23:17.830151 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:17 crc kubenswrapper[4966]: E1217 08:23:17.830313 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:17 crc kubenswrapper[4966]: I1217 08:23:17.830177 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:17 crc kubenswrapper[4966]: E1217 08:23:17.830438 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:17 crc kubenswrapper[4966]: E1217 08:23:17.830614 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:17 crc kubenswrapper[4966]: E1217 08:23:17.830764 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:19 crc kubenswrapper[4966]: I1217 08:23:19.830309 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:19 crc kubenswrapper[4966]: I1217 08:23:19.830372 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:19 crc kubenswrapper[4966]: I1217 08:23:19.830372 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:19 crc kubenswrapper[4966]: E1217 08:23:19.830468 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:19 crc kubenswrapper[4966]: I1217 08:23:19.830638 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:19 crc kubenswrapper[4966]: E1217 08:23:19.830706 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:19 crc kubenswrapper[4966]: E1217 08:23:19.830609 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:19 crc kubenswrapper[4966]: E1217 08:23:19.830796 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:19 crc kubenswrapper[4966]: E1217 08:23:19.903136 4966 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 17 08:23:21 crc kubenswrapper[4966]: I1217 08:23:21.829740 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:21 crc kubenswrapper[4966]: I1217 08:23:21.829816 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:21 crc kubenswrapper[4966]: I1217 08:23:21.829912 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:21 crc kubenswrapper[4966]: E1217 08:23:21.830022 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:21 crc kubenswrapper[4966]: I1217 08:23:21.830054 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:21 crc kubenswrapper[4966]: E1217 08:23:21.830196 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:21 crc kubenswrapper[4966]: E1217 08:23:21.830362 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:21 crc kubenswrapper[4966]: E1217 08:23:21.830407 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:22 crc kubenswrapper[4966]: I1217 08:23:22.831338 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:23:23 crc kubenswrapper[4966]: I1217 08:23:23.494020 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/3.log" Dec 17 08:23:23 crc kubenswrapper[4966]: I1217 08:23:23.496775 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerStarted","Data":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} Dec 17 08:23:23 crc kubenswrapper[4966]: I1217 08:23:23.497206 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:23:23 crc kubenswrapper[4966]: I1217 08:23:23.830133 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:23 crc kubenswrapper[4966]: I1217 08:23:23.830178 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:23 crc kubenswrapper[4966]: E1217 08:23:23.830270 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:23 crc kubenswrapper[4966]: I1217 08:23:23.830461 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:23 crc kubenswrapper[4966]: E1217 08:23:23.830535 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:23 crc kubenswrapper[4966]: I1217 08:23:23.830665 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:23 crc kubenswrapper[4966]: E1217 08:23:23.830724 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:23 crc kubenswrapper[4966]: E1217 08:23:23.830888 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:23 crc kubenswrapper[4966]: I1217 08:23:23.840186 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podStartSLOduration=120.840162864 podStartE2EDuration="2m0.840162864s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:23.52482142 +0000 UTC m=+139.069891382" watchObservedRunningTime="2025-12-17 08:23:23.840162864 +0000 UTC m=+139.385232806" Dec 17 08:23:23 crc kubenswrapper[4966]: I1217 08:23:23.841136 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nqmws"] Dec 17 08:23:24 crc kubenswrapper[4966]: I1217 08:23:24.500045 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:24 crc kubenswrapper[4966]: E1217 08:23:24.500520 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:24 crc kubenswrapper[4966]: E1217 08:23:24.903785 4966 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 17 08:23:25 crc kubenswrapper[4966]: I1217 08:23:25.830130 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:25 crc kubenswrapper[4966]: I1217 08:23:25.830175 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:25 crc kubenswrapper[4966]: I1217 08:23:25.830207 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:25 crc kubenswrapper[4966]: I1217 08:23:25.830151 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:25 crc kubenswrapper[4966]: E1217 08:23:25.830272 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:25 crc kubenswrapper[4966]: E1217 08:23:25.830370 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:25 crc kubenswrapper[4966]: E1217 08:23:25.830476 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:25 crc kubenswrapper[4966]: E1217 08:23:25.830675 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:27 crc kubenswrapper[4966]: I1217 08:23:27.830273 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:27 crc kubenswrapper[4966]: I1217 08:23:27.830306 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:27 crc kubenswrapper[4966]: E1217 08:23:27.830713 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:27 crc kubenswrapper[4966]: I1217 08:23:27.830405 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:27 crc kubenswrapper[4966]: I1217 08:23:27.830343 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:27 crc kubenswrapper[4966]: E1217 08:23:27.830951 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:27 crc kubenswrapper[4966]: E1217 08:23:27.831076 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:27 crc kubenswrapper[4966]: E1217 08:23:27.831177 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:29 crc kubenswrapper[4966]: I1217 08:23:29.829618 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:29 crc kubenswrapper[4966]: I1217 08:23:29.829664 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:29 crc kubenswrapper[4966]: I1217 08:23:29.829624 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:29 crc kubenswrapper[4966]: E1217 08:23:29.829759 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:23:29 crc kubenswrapper[4966]: I1217 08:23:29.829857 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:29 crc kubenswrapper[4966]: E1217 08:23:29.829978 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:23:29 crc kubenswrapper[4966]: E1217 08:23:29.830076 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:23:29 crc kubenswrapper[4966]: E1217 08:23:29.830187 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nqmws" podUID="6cf617b8-bd09-49de-b188-9c86d5fafd57" Dec 17 08:23:30 crc kubenswrapper[4966]: I1217 08:23:30.616467 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.616781 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:25:32.616757469 +0000 UTC m=+268.161827431 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:30 crc kubenswrapper[4966]: I1217 08:23:30.717307 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:30 crc kubenswrapper[4966]: I1217 08:23:30.717358 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:30 crc kubenswrapper[4966]: I1217 08:23:30.717389 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:30 crc kubenswrapper[4966]: I1217 08:23:30.717415 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717497 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717513 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717523 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717544 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717559 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:25:32.717544204 +0000 UTC m=+268.262614146 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717561 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717595 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:25:32.717585925 +0000 UTC m=+268.262655867 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717560 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717631 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:25:32.717623896 +0000 UTC m=+268.262693838 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717531 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717649 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:23:30 crc kubenswrapper[4966]: E1217 08:23:30.717674 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:25:32.717666477 +0000 UTC m=+268.262736419 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.829315 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.829356 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.829362 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.829411 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.832047 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.832278 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.832355 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.832379 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.832414 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 17 08:23:31 crc kubenswrapper[4966]: I1217 08:23:31.833159 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.718298 4966 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.806025 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bnjl5"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.806619 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.808414 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.808828 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.810949 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.812001 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.813445 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-r5fmr"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.813947 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r5fmr" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.819061 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.819689 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.820150 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v6b7k"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.820631 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.821946 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-knt89"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.822509 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ljxtf"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.822835 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.822948 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.823481 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9pzvc"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.824424 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.842795 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.843298 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f4bsv"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.843693 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.847480 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.847486 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.853582 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.854141 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.854323 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.854331 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.854331 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.854979 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.855076 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.855182 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.855255 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.855610 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.856372 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-b998v"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.856831 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.857654 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.859638 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.859949 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.863511 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.863808 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.863621 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.864205 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.864333 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.864444 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.864500 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.865198 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.865439 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.869227 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.869377 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.869425 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.869527 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.869556 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.869644 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.873842 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.874176 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.874725 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.874809 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.875256 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.876469 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.876797 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.876832 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.877133 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.877677 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.878084 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.879288 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.882461 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.882947 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.892237 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.892423 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.892506 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.899956 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.900087 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.900163 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.900265 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.900318 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.900364 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.900416 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.900484 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.900541 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.900597 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.903966 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-922bk"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.904105 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.904652 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.904707 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.904720 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.905213 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.905289 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.905413 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.906652 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.907210 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-k5zzh"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.907661 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.907861 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.908549 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.909233 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.916905 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dae6e8c2-ac2d-479b-83c1-c10aff34a548-config\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.916939 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-oauth-serving-cert\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.916963 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-etcd-client\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.916980 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-config\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.916996 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-image-import-ca\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917010 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbzqw\" (UniqueName: \"kubernetes.io/projected/41dbc543-cace-4c8f-8db3-e07b984c0919-kube-api-access-sbzqw\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917025 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bff81677-e3e6-4e74-81ba-d89156398d45-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917042 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917058 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jslj9\" (UniqueName: \"kubernetes.io/projected/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-kube-api-access-jslj9\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917083 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/906afeba-a01c-4762-921a-35f608da53b2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-75n4w\" (UID: \"906afeba-a01c-4762-921a-35f608da53b2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917100 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5cc70d0-a401-46b1-b102-a397fa0183c7-config\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917116 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b053b6a3-d480-4dee-9c7f-146c7d084e06-config\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917289 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917351 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gfkj\" (UniqueName: \"kubernetes.io/projected/e0296a9b-6991-4eb3-81d4-b7c710bc9acf-kube-api-access-7gfkj\") pod \"openshift-apiserver-operator-796bbdcf4f-phjzd\" (UID: \"e0296a9b-6991-4eb3-81d4-b7c710bc9acf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917374 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-node-pullsecrets\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917391 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b053b6a3-d480-4dee-9c7f-146c7d084e06-serving-cert\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917408 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-service-ca\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917426 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bff81677-e3e6-4e74-81ba-d89156398d45-images\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917440 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcqx5\" (UniqueName: \"kubernetes.io/projected/992684e8-e50b-44e7-b30c-7169f4877695-kube-api-access-gcqx5\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917456 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dae6e8c2-ac2d-479b-83c1-c10aff34a548-etcd-service-ca\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917472 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-serving-cert\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917488 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vv8f\" (UniqueName: \"kubernetes.io/projected/7a83b3e0-2092-491a-8844-1aec20fb2752-kube-api-access-6vv8f\") pod \"openshift-config-operator-7777fb866f-knt89\" (UID: \"7a83b3e0-2092-491a-8844-1aec20fb2752\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917505 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0296a9b-6991-4eb3-81d4-b7c710bc9acf-config\") pod \"openshift-apiserver-operator-796bbdcf4f-phjzd\" (UID: \"e0296a9b-6991-4eb3-81d4-b7c710bc9acf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917520 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dae6e8c2-ac2d-479b-83c1-c10aff34a548-serving-cert\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917547 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-oauth-config\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917590 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a83b3e0-2092-491a-8844-1aec20fb2752-serving-cert\") pod \"openshift-config-operator-7777fb866f-knt89\" (UID: \"7a83b3e0-2092-491a-8844-1aec20fb2752\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917653 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-config\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917727 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-serving-cert\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917753 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dae6e8c2-ac2d-479b-83c1-c10aff34a548-etcd-client\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.917792 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k9rg\" (UniqueName: \"kubernetes.io/projected/50255104-8dc1-485a-b18b-b2f86d9150b4-kube-api-access-6k9rg\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918138 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dae6e8c2-ac2d-479b-83c1-c10aff34a548-etcd-ca\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918176 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a5cc70d0-a401-46b1-b102-a397fa0183c7-auth-proxy-config\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918198 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50255104-8dc1-485a-b18b-b2f86d9150b4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918219 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-client-ca\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918240 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-545cz\" (UniqueName: \"kubernetes.io/projected/1b60ed2c-77ac-4cbc-9b57-48ed79419073-kube-api-access-545cz\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918261 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-trusted-ca-bundle\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918281 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0296a9b-6991-4eb3-81d4-b7c710bc9acf-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-phjzd\" (UID: \"e0296a9b-6991-4eb3-81d4-b7c710bc9acf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918302 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-encryption-config\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918324 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50255104-8dc1-485a-b18b-b2f86d9150b4-config\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918345 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-config\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918373 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-console-config\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918408 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bff81677-e3e6-4e74-81ba-d89156398d45-config\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918430 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-etcd-serving-ca\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918459 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50255104-8dc1-485a-b18b-b2f86d9150b4-service-ca-bundle\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918493 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-audit\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918514 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v42qm\" (UniqueName: \"kubernetes.io/projected/a5cc70d0-a401-46b1-b102-a397fa0183c7-kube-api-access-v42qm\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918532 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-trusted-ca-bundle\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918550 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlw97\" (UniqueName: \"kubernetes.io/projected/b053b6a3-d480-4dee-9c7f-146c7d084e06-kube-api-access-tlw97\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918587 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-audit-dir\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918605 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41dbc543-cace-4c8f-8db3-e07b984c0919-serving-cert\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918633 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q98x6\" (UniqueName: \"kubernetes.io/projected/906afeba-a01c-4762-921a-35f608da53b2-kube-api-access-q98x6\") pod \"cluster-samples-operator-665b6dd947-75n4w\" (UID: \"906afeba-a01c-4762-921a-35f608da53b2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918656 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm7hk\" (UniqueName: \"kubernetes.io/projected/091f07a2-acd7-4d1e-b242-220c47716c25-kube-api-access-rm7hk\") pod \"downloads-7954f5f757-r5fmr\" (UID: \"091f07a2-acd7-4d1e-b242-220c47716c25\") " pod="openshift-console/downloads-7954f5f757-r5fmr" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918676 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c87kj\" (UniqueName: \"kubernetes.io/projected/bff81677-e3e6-4e74-81ba-d89156398d45-kube-api-access-c87kj\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918708 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50255104-8dc1-485a-b18b-b2f86d9150b4-serving-cert\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918732 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-client-ca\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918755 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7a83b3e0-2092-491a-8844-1aec20fb2752-available-featuregates\") pod \"openshift-config-operator-7777fb866f-knt89\" (UID: \"7a83b3e0-2092-491a-8844-1aec20fb2752\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918775 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b60ed2c-77ac-4cbc-9b57-48ed79419073-serving-cert\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918797 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b053b6a3-d480-4dee-9c7f-146c7d084e06-trusted-ca\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918918 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmbgg\" (UniqueName: \"kubernetes.io/projected/dae6e8c2-ac2d-479b-83c1-c10aff34a548-kube-api-access-bmbgg\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.918947 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a5cc70d0-a401-46b1-b102-a397fa0183c7-machine-approver-tls\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.919399 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.924121 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.928043 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.928518 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.928802 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.929293 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.929702 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.930038 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.931802 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxfxg"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.932222 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.932463 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.933588 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.934006 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.936656 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.947645 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sfbnp"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.948615 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.949528 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.940675 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.950015 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.950085 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.955299 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.960505 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.960865 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.967253 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.967381 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-587gm"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.967441 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.967260 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.941305 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.945078 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.945169 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.945510 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.951433 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.968329 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.968642 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.952734 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.972175 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.977621 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.940925 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.955042 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.955174 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.980386 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.980769 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.981195 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.981652 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.982076 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.983021 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.983363 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.983481 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.983601 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.953706 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.983840 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.984074 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.984385 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.984438 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.984571 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.984676 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.984976 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.992650 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.994033 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.994061 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.994709 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.994838 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.994990 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.996298 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.996428 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-clm56"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.997201 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.998176 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q"] Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.998761 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:39 crc kubenswrapper[4966]: I1217 08:23:39.999808 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.000445 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.006452 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.006775 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.007501 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.017849 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.018629 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.019588 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0296a9b-6991-4eb3-81d4-b7c710bc9acf-config\") pod \"openshift-apiserver-operator-796bbdcf4f-phjzd\" (UID: \"e0296a9b-6991-4eb3-81d4-b7c710bc9acf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020350 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dae6e8c2-ac2d-479b-83c1-c10aff34a548-serving-cert\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020463 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vv8f\" (UniqueName: \"kubernetes.io/projected/7a83b3e0-2092-491a-8844-1aec20fb2752-kube-api-access-6vv8f\") pod \"openshift-config-operator-7777fb866f-knt89\" (UID: \"7a83b3e0-2092-491a-8844-1aec20fb2752\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020586 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcf98\" (UniqueName: \"kubernetes.io/projected/fadc152c-c5d3-4cec-b2bb-c35126687c76-kube-api-access-rcf98\") pod \"kube-storage-version-migrator-operator-b67b599dd-8gpfd\" (UID: \"fadc152c-c5d3-4cec-b2bb-c35126687c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020708 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a83b3e0-2092-491a-8844-1aec20fb2752-serving-cert\") pod \"openshift-config-operator-7777fb866f-knt89\" (UID: \"7a83b3e0-2092-491a-8844-1aec20fb2752\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020760 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020785 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-oauth-config\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020816 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-config\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020858 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-serving-cert\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020928 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dae6e8c2-ac2d-479b-83c1-c10aff34a548-etcd-client\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020950 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4n9j\" (UniqueName: \"kubernetes.io/projected/021d4c31-6de9-4e76-ac80-0f345b3d7f81-kube-api-access-q4n9j\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.020991 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a5cc70d0-a401-46b1-b102-a397fa0183c7-auth-proxy-config\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021015 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50255104-8dc1-485a-b18b-b2f86d9150b4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021035 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k9rg\" (UniqueName: \"kubernetes.io/projected/50255104-8dc1-485a-b18b-b2f86d9150b4-kube-api-access-6k9rg\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021075 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dae6e8c2-ac2d-479b-83c1-c10aff34a548-etcd-ca\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021099 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021123 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-client-ca\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021167 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-545cz\" (UniqueName: \"kubernetes.io/projected/1b60ed2c-77ac-4cbc-9b57-48ed79419073-kube-api-access-545cz\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021191 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021233 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0296a9b-6991-4eb3-81d4-b7c710bc9acf-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-phjzd\" (UID: \"e0296a9b-6991-4eb3-81d4-b7c710bc9acf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021255 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-encryption-config\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021275 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50255104-8dc1-485a-b18b-b2f86d9150b4-config\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021317 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-trusted-ca-bundle\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021341 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88d927f5-13ac-470d-8544-9a4aa1dc55ff-config\") pod \"kube-apiserver-operator-766d6c64bb-429md\" (UID: \"88d927f5-13ac-470d-8544-9a4aa1dc55ff\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021366 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bff81677-e3e6-4e74-81ba-d89156398d45-config\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021415 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-config\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021437 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-console-config\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021486 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-audit\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021510 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-etcd-serving-ca\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021532 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50255104-8dc1-485a-b18b-b2f86d9150b4-service-ca-bundle\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021588 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b1849b5-04d6-4550-bd09-0f69c712b283-service-ca-bundle\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021652 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v42qm\" (UniqueName: \"kubernetes.io/projected/a5cc70d0-a401-46b1-b102-a397fa0183c7-kube-api-access-v42qm\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021677 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-trusted-ca-bundle\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021724 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlw97\" (UniqueName: \"kubernetes.io/projected/b053b6a3-d480-4dee-9c7f-146c7d084e06-kube-api-access-tlw97\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021770 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8b1849b5-04d6-4550-bd09-0f69c712b283-stats-auth\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021821 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-audit-dir\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021932 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41dbc543-cace-4c8f-8db3-e07b984c0919-serving-cert\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.021961 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-dir\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022232 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88d927f5-13ac-470d-8544-9a4aa1dc55ff-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-429md\" (UID: \"88d927f5-13ac-470d-8544-9a4aa1dc55ff\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022305 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf9rh\" (UniqueName: \"kubernetes.io/projected/5ab4a948-e08e-4c50-a900-cfab1e8b7d4f-kube-api-access-wf9rh\") pod \"dns-operator-744455d44c-k5zzh\" (UID: \"5ab4a948-e08e-4c50-a900-cfab1e8b7d4f\") " pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022334 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjlrc\" (UniqueName: \"kubernetes.io/projected/8b1849b5-04d6-4550-bd09-0f69c712b283-kube-api-access-kjlrc\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022359 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022412 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm7hk\" (UniqueName: \"kubernetes.io/projected/091f07a2-acd7-4d1e-b242-220c47716c25-kube-api-access-rm7hk\") pod \"downloads-7954f5f757-r5fmr\" (UID: \"091f07a2-acd7-4d1e-b242-220c47716c25\") " pod="openshift-console/downloads-7954f5f757-r5fmr" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022436 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c87kj\" (UniqueName: \"kubernetes.io/projected/bff81677-e3e6-4e74-81ba-d89156398d45-kube-api-access-c87kj\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022483 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q98x6\" (UniqueName: \"kubernetes.io/projected/906afeba-a01c-4762-921a-35f608da53b2-kube-api-access-q98x6\") pod \"cluster-samples-operator-665b6dd947-75n4w\" (UID: \"906afeba-a01c-4762-921a-35f608da53b2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022510 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50255104-8dc1-485a-b18b-b2f86d9150b4-serving-cert\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022560 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-policies\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022586 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022611 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fadc152c-c5d3-4cec-b2bb-c35126687c76-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8gpfd\" (UID: \"fadc152c-c5d3-4cec-b2bb-c35126687c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022675 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-client-ca\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022703 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7a83b3e0-2092-491a-8844-1aec20fb2752-available-featuregates\") pod \"openshift-config-operator-7777fb866f-knt89\" (UID: \"7a83b3e0-2092-491a-8844-1aec20fb2752\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022753 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b1849b5-04d6-4550-bd09-0f69c712b283-metrics-certs\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022774 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88d927f5-13ac-470d-8544-9a4aa1dc55ff-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-429md\" (UID: \"88d927f5-13ac-470d-8544-9a4aa1dc55ff\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022816 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b053b6a3-d480-4dee-9c7f-146c7d084e06-trusted-ca\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022843 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b60ed2c-77ac-4cbc-9b57-48ed79419073-serving-cert\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022866 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022921 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmbgg\" (UniqueName: \"kubernetes.io/projected/dae6e8c2-ac2d-479b-83c1-c10aff34a548-kube-api-access-bmbgg\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022947 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a5cc70d0-a401-46b1-b102-a397fa0183c7-machine-approver-tls\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.022993 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ab4a948-e08e-4c50-a900-cfab1e8b7d4f-metrics-tls\") pod \"dns-operator-744455d44c-k5zzh\" (UID: \"5ab4a948-e08e-4c50-a900-cfab1e8b7d4f\") " pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023028 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-etcd-client\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023072 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dae6e8c2-ac2d-479b-83c1-c10aff34a548-config\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023096 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-oauth-serving-cert\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023142 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/eb93a1c3-72e1-4fe1-96c3-742263dced3e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-clm56\" (UID: \"eb93a1c3-72e1-4fe1-96c3-742263dced3e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023171 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-config\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023196 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-image-import-ca\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023240 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbzqw\" (UniqueName: \"kubernetes.io/projected/41dbc543-cace-4c8f-8db3-e07b984c0919-kube-api-access-sbzqw\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023264 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023311 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023340 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bff81677-e3e6-4e74-81ba-d89156398d45-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023362 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023409 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023436 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5cc70d0-a401-46b1-b102-a397fa0183c7-config\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023480 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jslj9\" (UniqueName: \"kubernetes.io/projected/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-kube-api-access-jslj9\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023504 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/906afeba-a01c-4762-921a-35f608da53b2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-75n4w\" (UID: \"906afeba-a01c-4762-921a-35f608da53b2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023553 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b053b6a3-d480-4dee-9c7f-146c7d084e06-config\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023580 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fadc152c-c5d3-4cec-b2bb-c35126687c76-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8gpfd\" (UID: \"fadc152c-c5d3-4cec-b2bb-c35126687c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023606 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023654 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gfkj\" (UniqueName: \"kubernetes.io/projected/e0296a9b-6991-4eb3-81d4-b7c710bc9acf-kube-api-access-7gfkj\") pod \"openshift-apiserver-operator-796bbdcf4f-phjzd\" (UID: \"e0296a9b-6991-4eb3-81d4-b7c710bc9acf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023682 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-node-pullsecrets\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023727 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b053b6a3-d480-4dee-9c7f-146c7d084e06-serving-cert\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023749 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-service-ca\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023815 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bff81677-e3e6-4e74-81ba-d89156398d45-images\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023842 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcqx5\" (UniqueName: \"kubernetes.io/projected/992684e8-e50b-44e7-b30c-7169f4877695-kube-api-access-gcqx5\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023897 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8b1849b5-04d6-4550-bd09-0f69c712b283-default-certificate\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023925 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.023976 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dae6e8c2-ac2d-479b-83c1-c10aff34a548-etcd-service-ca\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.024001 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-serving-cert\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.024046 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgqk5\" (UniqueName: \"kubernetes.io/projected/eb93a1c3-72e1-4fe1-96c3-742263dced3e-kube-api-access-tgqk5\") pod \"multus-admission-controller-857f4d67dd-clm56\" (UID: \"eb93a1c3-72e1-4fe1-96c3-742263dced3e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.026556 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7a83b3e0-2092-491a-8844-1aec20fb2752-available-featuregates\") pod \"openshift-config-operator-7777fb866f-knt89\" (UID: \"7a83b3e0-2092-491a-8844-1aec20fb2752\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.026765 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dae6e8c2-ac2d-479b-83c1-c10aff34a548-etcd-ca\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.027470 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-audit-dir\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.028277 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-trusted-ca-bundle\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.028582 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-client-ca\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.032129 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5cc70d0-a401-46b1-b102-a397fa0183c7-config\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.032217 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.032702 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.033569 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.041377 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b053b6a3-d480-4dee-9c7f-146c7d084e06-config\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.038061 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0296a9b-6991-4eb3-81d4-b7c710bc9acf-config\") pod \"openshift-apiserver-operator-796bbdcf4f-phjzd\" (UID: \"e0296a9b-6991-4eb3-81d4-b7c710bc9acf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.038633 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.041786 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-node-pullsecrets\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.040688 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b053b6a3-d480-4dee-9c7f-146c7d084e06-trusted-ca\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.041015 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-config\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.042287 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.042449 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0296a9b-6991-4eb3-81d4-b7c710bc9acf-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-phjzd\" (UID: \"e0296a9b-6991-4eb3-81d4-b7c710bc9acf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.037736 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.043595 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/906afeba-a01c-4762-921a-35f608da53b2-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-75n4w\" (UID: \"906afeba-a01c-4762-921a-35f608da53b2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.046062 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-audit\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.046299 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50255104-8dc1-485a-b18b-b2f86d9150b4-config\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.046732 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-etcd-serving-ca\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.046960 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bff81677-e3e6-4e74-81ba-d89156398d45-config\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.047559 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a5cc70d0-a401-46b1-b102-a397fa0183c7-auth-proxy-config\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.048691 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bff81677-e3e6-4e74-81ba-d89156398d45-images\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.049199 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dae6e8c2-ac2d-479b-83c1-c10aff34a548-etcd-service-ca\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.049573 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50255104-8dc1-485a-b18b-b2f86d9150b4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.049701 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b60ed2c-77ac-4cbc-9b57-48ed79419073-serving-cert\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.039169 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dae6e8c2-ac2d-479b-83c1-c10aff34a548-serving-cert\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.050031 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a83b3e0-2092-491a-8844-1aec20fb2752-serving-cert\") pod \"openshift-config-operator-7777fb866f-knt89\" (UID: \"7a83b3e0-2092-491a-8844-1aec20fb2752\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.052486 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a5cc70d0-a401-46b1-b102-a397fa0183c7-machine-approver-tls\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.052671 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-config\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.052902 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-service-ca\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.053010 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41dbc543-cace-4c8f-8db3-e07b984c0919-serving-cert\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.053085 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.053308 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-console-config\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.053401 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50255104-8dc1-485a-b18b-b2f86d9150b4-service-ca-bundle\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.053400 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b053b6a3-d480-4dee-9c7f-146c7d084e06-serving-cert\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.053654 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-km2tl"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.054154 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r5fmr"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.054236 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.054290 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-oauth-serving-cert\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.054477 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.054517 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-trusted-ca-bundle\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.055028 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-config\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.055379 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-image-import-ca\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.055559 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dae6e8c2-ac2d-479b-83c1-c10aff34a548-config\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.056284 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-oauth-config\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.057214 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-etcd-client\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.059442 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-client-ca\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.059839 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.066533 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.066469 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bnjl5"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.060822 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.060451 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dae6e8c2-ac2d-479b-83c1-c10aff34a548-etcd-client\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.060697 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-encryption-config\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.068749 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.070125 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.071231 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.072245 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-serving-cert\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.072489 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8j46g"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.073142 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.074205 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.078056 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-serving-cert\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.078268 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-knt89"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.078249 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50255104-8dc1-485a-b18b-b2f86d9150b4-serving-cert\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.086229 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bff81677-e3e6-4e74-81ba-d89156398d45-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.089948 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-b998v"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.094205 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-clm56"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.096002 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-k5zzh"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.099936 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ljxtf"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.100698 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.105809 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9pzvc"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.108183 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.109416 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v6b7k"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.110110 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-9f9j7"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.110361 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.111225 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.112058 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.113557 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-922bk"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.115402 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-6r6bw"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.115828 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6r6bw" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.116935 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f4bsv"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.118060 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.119227 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.121055 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lcqdh"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.122398 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.122882 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.124284 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.124804 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgqk5\" (UniqueName: \"kubernetes.io/projected/eb93a1c3-72e1-4fe1-96c3-742263dced3e-kube-api-access-tgqk5\") pod \"multus-admission-controller-857f4d67dd-clm56\" (UID: \"eb93a1c3-72e1-4fe1-96c3-742263dced3e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.125020 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcf98\" (UniqueName: \"kubernetes.io/projected/fadc152c-c5d3-4cec-b2bb-c35126687c76-kube-api-access-rcf98\") pod \"kube-storage-version-migrator-operator-b67b599dd-8gpfd\" (UID: \"fadc152c-c5d3-4cec-b2bb-c35126687c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.125169 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.125495 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4n9j\" (UniqueName: \"kubernetes.io/projected/021d4c31-6de9-4e76-ac80-0f345b3d7f81-kube-api-access-q4n9j\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.125659 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.125739 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.126009 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.126147 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88d927f5-13ac-470d-8544-9a4aa1dc55ff-config\") pod \"kube-apiserver-operator-766d6c64bb-429md\" (UID: \"88d927f5-13ac-470d-8544-9a4aa1dc55ff\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.126805 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b1849b5-04d6-4550-bd09-0f69c712b283-service-ca-bundle\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.127104 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88d927f5-13ac-470d-8544-9a4aa1dc55ff-config\") pod \"kube-apiserver-operator-766d6c64bb-429md\" (UID: \"88d927f5-13ac-470d-8544-9a4aa1dc55ff\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.127498 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8b1849b5-04d6-4550-bd09-0f69c712b283-stats-auth\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.127762 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-dir\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.128165 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88d927f5-13ac-470d-8544-9a4aa1dc55ff-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-429md\" (UID: \"88d927f5-13ac-470d-8544-9a4aa1dc55ff\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.129157 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf9rh\" (UniqueName: \"kubernetes.io/projected/5ab4a948-e08e-4c50-a900-cfab1e8b7d4f-kube-api-access-wf9rh\") pod \"dns-operator-744455d44c-k5zzh\" (UID: \"5ab4a948-e08e-4c50-a900-cfab1e8b7d4f\") " pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.129412 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjlrc\" (UniqueName: \"kubernetes.io/projected/8b1849b5-04d6-4550-bd09-0f69c712b283-kube-api-access-kjlrc\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.131421 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.128822 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.128115 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-dir\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.127715 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.132103 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.132155 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.128786 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.131355 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88d927f5-13ac-470d-8544-9a4aa1dc55ff-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-429md\" (UID: \"88d927f5-13ac-470d-8544-9a4aa1dc55ff\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.132690 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-policies\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.132807 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.132968 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fadc152c-c5d3-4cec-b2bb-c35126687c76-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8gpfd\" (UID: \"fadc152c-c5d3-4cec-b2bb-c35126687c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133092 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b1849b5-04d6-4550-bd09-0f69c712b283-metrics-certs\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133195 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88d927f5-13ac-470d-8544-9a4aa1dc55ff-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-429md\" (UID: \"88d927f5-13ac-470d-8544-9a4aa1dc55ff\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133296 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133413 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ab4a948-e08e-4c50-a900-cfab1e8b7d4f-metrics-tls\") pod \"dns-operator-744455d44c-k5zzh\" (UID: \"5ab4a948-e08e-4c50-a900-cfab1e8b7d4f\") " pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133520 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/eb93a1c3-72e1-4fe1-96c3-742263dced3e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-clm56\" (UID: \"eb93a1c3-72e1-4fe1-96c3-742263dced3e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133623 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133716 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133828 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133977 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.134085 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fadc152c-c5d3-4cec-b2bb-c35126687c76-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8gpfd\" (UID: \"fadc152c-c5d3-4cec-b2bb-c35126687c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.134207 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8b1849b5-04d6-4550-bd09-0f69c712b283-default-certificate\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.134308 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.133151 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.135205 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.136976 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.137590 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-policies\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.137937 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.138144 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.138257 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.138993 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.139098 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.139465 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.140046 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sfbnp"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.140452 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.141644 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.143470 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.144567 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.145590 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.146593 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.147616 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8j46g"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.148668 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.149357 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9f9j7"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.150784 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lcqdh"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.151722 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxfxg"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.152836 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8rncq"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.154010 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-km2tl"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.154104 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.155626 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.156645 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.157105 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6r6bw"] Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.157535 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ab4a948-e08e-4c50-a900-cfab1e8b7d4f-metrics-tls\") pod \"dns-operator-744455d44c-k5zzh\" (UID: \"5ab4a948-e08e-4c50-a900-cfab1e8b7d4f\") " pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.170146 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.176203 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.195694 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.204201 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.209020 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.233732 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.248495 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.268366 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.288596 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.308170 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.328153 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.348202 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.369367 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.388585 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.409150 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.429364 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.448265 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.468973 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.489616 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.508630 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.529112 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.544346 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8b1849b5-04d6-4550-bd09-0f69c712b283-stats-auth\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.549773 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.558075 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b1849b5-04d6-4550-bd09-0f69c712b283-metrics-certs\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.570113 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.589372 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.600369 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8b1849b5-04d6-4550-bd09-0f69c712b283-default-certificate\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.609115 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.630192 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.638192 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b1849b5-04d6-4550-bd09-0f69c712b283-service-ca-bundle\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.649306 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.669608 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.679297 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fadc152c-c5d3-4cec-b2bb-c35126687c76-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8gpfd\" (UID: \"fadc152c-c5d3-4cec-b2bb-c35126687c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.688831 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.694730 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fadc152c-c5d3-4cec-b2bb-c35126687c76-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8gpfd\" (UID: \"fadc152c-c5d3-4cec-b2bb-c35126687c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.710062 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.729209 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.749600 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.789098 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.809228 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.829375 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.849325 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.869217 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.889418 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.899955 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/eb93a1c3-72e1-4fe1-96c3-742263dced3e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-clm56\" (UID: \"eb93a1c3-72e1-4fe1-96c3-742263dced3e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.909479 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.929361 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.949401 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.968475 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 17 08:23:40 crc kubenswrapper[4966]: I1217 08:23:40.989303 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.009829 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.027476 4966 request.go:700] Waited for 1.018916914s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.049066 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.068698 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.090330 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.108582 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.130325 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.170557 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vv8f\" (UniqueName: \"kubernetes.io/projected/7a83b3e0-2092-491a-8844-1aec20fb2752-kube-api-access-6vv8f\") pod \"openshift-config-operator-7777fb866f-knt89\" (UID: \"7a83b3e0-2092-491a-8844-1aec20fb2752\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.194425 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k9rg\" (UniqueName: \"kubernetes.io/projected/50255104-8dc1-485a-b18b-b2f86d9150b4-kube-api-access-6k9rg\") pod \"authentication-operator-69f744f599-ljxtf\" (UID: \"50255104-8dc1-485a-b18b-b2f86d9150b4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.211094 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlw97\" (UniqueName: \"kubernetes.io/projected/b053b6a3-d480-4dee-9c7f-146c7d084e06-kube-api-access-tlw97\") pod \"console-operator-58897d9998-f4bsv\" (UID: \"b053b6a3-d480-4dee-9c7f-146c7d084e06\") " pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.228820 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-545cz\" (UniqueName: \"kubernetes.io/projected/1b60ed2c-77ac-4cbc-9b57-48ed79419073-kube-api-access-545cz\") pod \"route-controller-manager-6576b87f9c-gzzdt\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.254100 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jslj9\" (UniqueName: \"kubernetes.io/projected/c7bbf1fa-fe65-456b-97ff-c57e2786ad73-kube-api-access-jslj9\") pod \"apiserver-76f77b778f-922bk\" (UID: \"c7bbf1fa-fe65-456b-97ff-c57e2786ad73\") " pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.271800 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm7hk\" (UniqueName: \"kubernetes.io/projected/091f07a2-acd7-4d1e-b242-220c47716c25-kube-api-access-rm7hk\") pod \"downloads-7954f5f757-r5fmr\" (UID: \"091f07a2-acd7-4d1e-b242-220c47716c25\") " pod="openshift-console/downloads-7954f5f757-r5fmr" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.287922 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c87kj\" (UniqueName: \"kubernetes.io/projected/bff81677-e3e6-4e74-81ba-d89156398d45-kube-api-access-c87kj\") pod \"machine-api-operator-5694c8668f-bnjl5\" (UID: \"bff81677-e3e6-4e74-81ba-d89156398d45\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.306675 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q98x6\" (UniqueName: \"kubernetes.io/projected/906afeba-a01c-4762-921a-35f608da53b2-kube-api-access-q98x6\") pod \"cluster-samples-operator-665b6dd947-75n4w\" (UID: \"906afeba-a01c-4762-921a-35f608da53b2\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.321796 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gfkj\" (UniqueName: \"kubernetes.io/projected/e0296a9b-6991-4eb3-81d4-b7c710bc9acf-kube-api-access-7gfkj\") pod \"openshift-apiserver-operator-796bbdcf4f-phjzd\" (UID: \"e0296a9b-6991-4eb3-81d4-b7c710bc9acf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.328396 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.328435 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.344297 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.349910 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.353620 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r5fmr" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.364864 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.368845 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.389149 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.412421 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.429519 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.431389 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.443933 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.451798 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.470292 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.488933 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.491412 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.503599 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.512673 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.539419 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.571992 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmbgg\" (UniqueName: \"kubernetes.io/projected/dae6e8c2-ac2d-479b-83c1-c10aff34a548-kube-api-access-bmbgg\") pod \"etcd-operator-b45778765-9pzvc\" (UID: \"dae6e8c2-ac2d-479b-83c1-c10aff34a548\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.582319 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcqx5\" (UniqueName: \"kubernetes.io/projected/992684e8-e50b-44e7-b30c-7169f4877695-kube-api-access-gcqx5\") pod \"console-f9d7485db-b998v\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.589305 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v42qm\" (UniqueName: \"kubernetes.io/projected/a5cc70d0-a401-46b1-b102-a397fa0183c7-kube-api-access-v42qm\") pod \"machine-approver-56656f9798-j6vwq\" (UID: \"a5cc70d0-a401-46b1-b102-a397fa0183c7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.604926 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbzqw\" (UniqueName: \"kubernetes.io/projected/41dbc543-cace-4c8f-8db3-e07b984c0919-kube-api-access-sbzqw\") pod \"controller-manager-879f6c89f-v6b7k\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.610864 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.637322 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.644307 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.652248 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.668241 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.670387 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.705546 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.705605 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.709185 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.732050 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.750249 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.755886 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.770615 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.791423 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.801556 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd"] Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.814339 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.815892 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r5fmr"] Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.829157 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.833479 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.835217 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt"] Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.856651 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.871132 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.894069 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.901892 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bnjl5"] Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.910276 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.932522 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.945277 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-knt89"] Dec 17 08:23:41 crc kubenswrapper[4966]: W1217 08:23:41.945400 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b60ed2c_77ac_4cbc_9b57_48ed79419073.slice/crio-41acd75f509ba5f4494933aef48f5fab95fc579e70b34fd723cc9abcbdb0333c WatchSource:0}: Error finding container 41acd75f509ba5f4494933aef48f5fab95fc579e70b34fd723cc9abcbdb0333c: Status 404 returned error can't find the container with id 41acd75f509ba5f4494933aef48f5fab95fc579e70b34fd723cc9abcbdb0333c Dec 17 08:23:41 crc kubenswrapper[4966]: W1217 08:23:41.947178 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbff81677_e3e6_4e74_81ba_d89156398d45.slice/crio-d4804fb38d86da0aa3c4b4db3626158f1985617cc9504b3d7fb2e3fa9adb2ef6 WatchSource:0}: Error finding container d4804fb38d86da0aa3c4b4db3626158f1985617cc9504b3d7fb2e3fa9adb2ef6: Status 404 returned error can't find the container with id d4804fb38d86da0aa3c4b4db3626158f1985617cc9504b3d7fb2e3fa9adb2ef6 Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.948547 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.971900 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 17 08:23:41 crc kubenswrapper[4966]: W1217 08:23:41.977342 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a83b3e0_2092_491a_8844_1aec20fb2752.slice/crio-110ffb4ac084c2d6c96213776007877812c7874c0b9a672fe1848bfa0904d7b6 WatchSource:0}: Error finding container 110ffb4ac084c2d6c96213776007877812c7874c0b9a672fe1848bfa0904d7b6: Status 404 returned error can't find the container with id 110ffb4ac084c2d6c96213776007877812c7874c0b9a672fe1848bfa0904d7b6 Dec 17 08:23:41 crc kubenswrapper[4966]: I1217 08:23:41.989023 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.008606 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.027984 4966 request.go:700] Waited for 1.905362695s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.029444 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.053474 4966 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.072725 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.107751 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgqk5\" (UniqueName: \"kubernetes.io/projected/eb93a1c3-72e1-4fe1-96c3-742263dced3e-kube-api-access-tgqk5\") pod \"multus-admission-controller-857f4d67dd-clm56\" (UID: \"eb93a1c3-72e1-4fe1-96c3-742263dced3e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.127125 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcf98\" (UniqueName: \"kubernetes.io/projected/fadc152c-c5d3-4cec-b2bb-c35126687c76-kube-api-access-rcf98\") pod \"kube-storage-version-migrator-operator-b67b599dd-8gpfd\" (UID: \"fadc152c-c5d3-4cec-b2bb-c35126687c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.159480 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4n9j\" (UniqueName: \"kubernetes.io/projected/021d4c31-6de9-4e76-ac80-0f345b3d7f81-kube-api-access-q4n9j\") pod \"oauth-openshift-558db77b4-nxfxg\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.175578 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf9rh\" (UniqueName: \"kubernetes.io/projected/5ab4a948-e08e-4c50-a900-cfab1e8b7d4f-kube-api-access-wf9rh\") pod \"dns-operator-744455d44c-k5zzh\" (UID: \"5ab4a948-e08e-4c50-a900-cfab1e8b7d4f\") " pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.183554 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjlrc\" (UniqueName: \"kubernetes.io/projected/8b1849b5-04d6-4550-bd09-0f69c712b283-kube-api-access-kjlrc\") pod \"router-default-5444994796-587gm\" (UID: \"8b1849b5-04d6-4550-bd09-0f69c712b283\") " pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.188753 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ljxtf"] Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.233243 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.236398 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.236738 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.238341 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88d927f5-13ac-470d-8544-9a4aa1dc55ff-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-429md\" (UID: \"88d927f5-13ac-470d-8544-9a4aa1dc55ff\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.253147 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.257445 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.275498 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-f4bsv"] Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.279512 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.283473 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w"] Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.287003 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.299744 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.315802 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-922bk"] Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368063 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b915a9bf-4d3d-44ae-aa8f-243283528d77-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zxqjj\" (UID: \"b915a9bf-4d3d-44ae-aa8f-243283528d77\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368114 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c6523c2a-3325-4334-ba30-7323b1597aaf-profile-collector-cert\") pod \"catalog-operator-68c6474976-f7sbq\" (UID: \"c6523c2a-3325-4334-ba30-7323b1597aaf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368152 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b915a9bf-4d3d-44ae-aa8f-243283528d77-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zxqjj\" (UID: \"b915a9bf-4d3d-44ae-aa8f-243283528d77\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368181 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b915a9bf-4d3d-44ae-aa8f-243283528d77-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zxqjj\" (UID: \"b915a9bf-4d3d-44ae-aa8f-243283528d77\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368209 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6db5f\" (UniqueName: \"kubernetes.io/projected/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-kube-api-access-6db5f\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368240 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmflq\" (UniqueName: \"kubernetes.io/projected/c3149310-e8c4-4c36-8ff6-993af0481122-kube-api-access-cmflq\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368290 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c6523c2a-3325-4334-ba30-7323b1597aaf-srv-cert\") pod \"catalog-operator-68c6474976-f7sbq\" (UID: \"c6523c2a-3325-4334-ba30-7323b1597aaf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368340 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c3149310-e8c4-4c36-8ff6-993af0481122-images\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368411 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368444 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d9189ca-6d00-4596-980a-690341335c2a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gw7bx\" (UID: \"4d9189ca-6d00-4596-980a-690341335c2a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368477 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368506 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-metrics-tls\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368533 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c3149310-e8c4-4c36-8ff6-993af0481122-proxy-tls\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368567 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-certificates\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368613 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368644 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f25b1b4a-7676-4f7c-847a-ee29217bb19d-webhook-cert\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368673 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c3149310-e8c4-4c36-8ff6-993af0481122-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368699 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd9mq\" (UniqueName: \"kubernetes.io/projected/15efc57c-51cd-465b-9f45-ab52ae521e55-kube-api-access-qd9mq\") pod \"openshift-controller-manager-operator-756b6f6bc6-nt2xl\" (UID: \"15efc57c-51cd-465b-9f45-ab52ae521e55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368727 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvjnc\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-kube-api-access-lvjnc\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368757 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15efc57c-51cd-465b-9f45-ab52ae521e55-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nt2xl\" (UID: \"15efc57c-51cd-465b-9f45-ab52ae521e55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368810 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368844 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-bound-sa-token\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368892 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpwpv\" (UniqueName: \"kubernetes.io/projected/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-kube-api-access-fpwpv\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368930 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f25b1b4a-7676-4f7c-847a-ee29217bb19d-apiservice-cert\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368962 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4dbs\" (UniqueName: \"kubernetes.io/projected/c6523c2a-3325-4334-ba30-7323b1597aaf-kube-api-access-v4dbs\") pod \"catalog-operator-68c6474976-f7sbq\" (UID: \"c6523c2a-3325-4334-ba30-7323b1597aaf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.368994 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-trusted-ca\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.369123 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-tls\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.369153 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.369202 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/569d9eb4-d037-4809-91a4-ee04d579cd21-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9q7zt\" (UID: \"569d9eb4-d037-4809-91a4-ee04d579cd21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.369235 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-897wc\" (UniqueName: \"kubernetes.io/projected/f25b1b4a-7676-4f7c-847a-ee29217bb19d-kube-api-access-897wc\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.369410 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: W1217 08:23:42.370683 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7bbf1fa_fe65_456b_97ff_c57e2786ad73.slice/crio-d42e3313755f8a89e8a85ff2ad0343b232a23ec8b3d0a032006d01c44ab30408 WatchSource:0}: Error finding container d42e3313755f8a89e8a85ff2ad0343b232a23ec8b3d0a032006d01c44ab30408: Status 404 returned error can't find the container with id d42e3313755f8a89e8a85ff2ad0343b232a23ec8b3d0a032006d01c44ab30408 Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.380784 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-trusted-ca\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.380823 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcpzl\" (UniqueName: \"kubernetes.io/projected/4d9189ca-6d00-4596-980a-690341335c2a-kube-api-access-xcpzl\") pod \"package-server-manager-789f6589d5-gw7bx\" (UID: \"4d9189ca-6d00-4596-980a-690341335c2a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.380849 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/569d9eb4-d037-4809-91a4-ee04d579cd21-config\") pod \"kube-controller-manager-operator-78b949d7b-9q7zt\" (UID: \"569d9eb4-d037-4809-91a4-ee04d579cd21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.380897 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/135afe15-d724-45ba-af8b-2a11862c594b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jqmsx\" (UID: \"135afe15-d724-45ba-af8b-2a11862c594b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.380974 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww9l2\" (UniqueName: \"kubernetes.io/projected/b6ac4609-63a9-400d-be84-cc6e4cb1a9cb-kube-api-access-ww9l2\") pod \"migrator-59844c95c7-c56mr\" (UID: \"b6ac4609-63a9-400d-be84-cc6e4cb1a9cb\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.380994 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f25b1b4a-7676-4f7c-847a-ee29217bb19d-tmpfs\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.381018 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrqzd\" (UniqueName: \"kubernetes.io/projected/135afe15-d724-45ba-af8b-2a11862c594b-kube-api-access-mrqzd\") pod \"control-plane-machine-set-operator-78cbb6b69f-jqmsx\" (UID: \"135afe15-d724-45ba-af8b-2a11862c594b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.381037 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: E1217 08:23:42.382396 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:42.88237433 +0000 UTC m=+158.427444272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.387931 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15efc57c-51cd-465b-9f45-ab52ae521e55-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nt2xl\" (UID: \"15efc57c-51cd-465b-9f45-ab52ae521e55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.388173 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/569d9eb4-d037-4809-91a4-ee04d579cd21-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9q7zt\" (UID: \"569d9eb4-d037-4809-91a4-ee04d579cd21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.403400 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v6b7k"] Dec 17 08:23:42 crc kubenswrapper[4966]: W1217 08:23:42.409029 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b1849b5_04d6_4550_bd09_0f69c712b283.slice/crio-a08f90b617b481adac64d1e2350620919db213d26efa142fb08fa950021fcb68 WatchSource:0}: Error finding container a08f90b617b481adac64d1e2350620919db213d26efa142fb08fa950021fcb68: Status 404 returned error can't find the container with id a08f90b617b481adac64d1e2350620919db213d26efa142fb08fa950021fcb68 Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.461167 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-b998v"] Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.493404 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.493893 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/569d9eb4-d037-4809-91a4-ee04d579cd21-config\") pod \"kube-controller-manager-operator-78b949d7b-9q7zt\" (UID: \"569d9eb4-d037-4809-91a4-ee04d579cd21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.494565 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/569d9eb4-d037-4809-91a4-ee04d579cd21-config\") pod \"kube-controller-manager-operator-78b949d7b-9q7zt\" (UID: \"569d9eb4-d037-4809-91a4-ee04d579cd21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: E1217 08:23:42.494639 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:42.993698758 +0000 UTC m=+158.538768730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.494704 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa-cert\") pod \"ingress-canary-6r6bw\" (UID: \"82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa\") " pod="openshift-ingress-canary/ingress-canary-6r6bw" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495264 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/04ea9c16-02a9-4116-9a09-63a19de322bd-certs\") pod \"machine-config-server-8rncq\" (UID: \"04ea9c16-02a9-4116-9a09-63a19de322bd\") " pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495296 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-976x6\" (UniqueName: \"kubernetes.io/projected/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-kube-api-access-976x6\") pod \"marketplace-operator-79b997595-km2tl\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495344 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/135afe15-d724-45ba-af8b-2a11862c594b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jqmsx\" (UID: \"135afe15-d724-45ba-af8b-2a11862c594b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495363 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww9l2\" (UniqueName: \"kubernetes.io/projected/b6ac4609-63a9-400d-be84-cc6e4cb1a9cb-kube-api-access-ww9l2\") pod \"migrator-59844c95c7-c56mr\" (UID: \"b6ac4609-63a9-400d-be84-cc6e4cb1a9cb\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495381 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-registration-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495422 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f25b1b4a-7676-4f7c-847a-ee29217bb19d-tmpfs\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495439 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d8c6\" (UniqueName: \"kubernetes.io/projected/aba2b97e-b9a0-4bc8-a480-84f64c2a0c55-kube-api-access-8d8c6\") pod \"machine-config-controller-84d6567774-lr2gg\" (UID: \"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495455 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-etcd-client\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495475 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thlgf\" (UniqueName: \"kubernetes.io/projected/97829faf-132d-43f9-8b6a-8b6e21c3f83e-kube-api-access-thlgf\") pod \"service-ca-9c57cc56f-8j46g\" (UID: \"97829faf-132d-43f9-8b6a-8b6e21c3f83e\") " pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495513 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-socket-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495532 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bc07aea-e7b5-473c-957c-a5295edc2f4e-config-volume\") pod \"collect-profiles-29432655-x4nw4\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495551 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrqzd\" (UniqueName: \"kubernetes.io/projected/135afe15-d724-45ba-af8b-2a11862c594b-kube-api-access-mrqzd\") pod \"control-plane-machine-set-operator-78cbb6b69f-jqmsx\" (UID: \"135afe15-d724-45ba-af8b-2a11862c594b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495575 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9hvw\" (UniqueName: \"kubernetes.io/projected/8410b9e5-a83f-4bef-a99a-736c634d9e41-kube-api-access-s9hvw\") pod \"dns-default-9f9j7\" (UID: \"8410b9e5-a83f-4bef-a99a-736c634d9e41\") " pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495653 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495684 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-km2tl\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495717 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/569d9eb4-d037-4809-91a4-ee04d579cd21-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9q7zt\" (UID: \"569d9eb4-d037-4809-91a4-ee04d579cd21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495743 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15efc57c-51cd-465b-9f45-ab52ae521e55-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nt2xl\" (UID: \"15efc57c-51cd-465b-9f45-ab52ae521e55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495779 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c6523c2a-3325-4334-ba30-7323b1597aaf-profile-collector-cert\") pod \"catalog-operator-68c6474976-f7sbq\" (UID: \"c6523c2a-3325-4334-ba30-7323b1597aaf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495800 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjknh\" (UniqueName: \"kubernetes.io/projected/9bc07aea-e7b5-473c-957c-a5295edc2f4e-kube-api-access-bjknh\") pod \"collect-profiles-29432655-x4nw4\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495834 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/32bd318f-c091-4913-889f-9097777ceb8e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jplds\" (UID: \"32bd318f-c091-4913-889f-9097777ceb8e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495861 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b915a9bf-4d3d-44ae-aa8f-243283528d77-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zxqjj\" (UID: \"b915a9bf-4d3d-44ae-aa8f-243283528d77\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.495977 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496003 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b915a9bf-4d3d-44ae-aa8f-243283528d77-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zxqjj\" (UID: \"b915a9bf-4d3d-44ae-aa8f-243283528d77\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496048 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/04ea9c16-02a9-4116-9a09-63a19de322bd-node-bootstrap-token\") pod \"machine-config-server-8rncq\" (UID: \"04ea9c16-02a9-4116-9a09-63a19de322bd\") " pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496184 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b915a9bf-4d3d-44ae-aa8f-243283528d77-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zxqjj\" (UID: \"b915a9bf-4d3d-44ae-aa8f-243283528d77\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496211 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6db5f\" (UniqueName: \"kubernetes.io/projected/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-kube-api-access-6db5f\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496233 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmflq\" (UniqueName: \"kubernetes.io/projected/c3149310-e8c4-4c36-8ff6-993af0481122-kube-api-access-cmflq\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496254 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-plugins-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496277 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c6523c2a-3325-4334-ba30-7323b1597aaf-srv-cert\") pod \"catalog-operator-68c6474976-f7sbq\" (UID: \"c6523c2a-3325-4334-ba30-7323b1597aaf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496298 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-audit-dir\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496345 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-mountpoint-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496368 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c3149310-e8c4-4c36-8ff6-993af0481122-images\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496415 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496457 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-audit-policies\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496480 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/399368ce-306d-4f0b-aef7-af19955056fa-serving-cert\") pod \"service-ca-operator-777779d784-sxhs2\" (UID: \"399368ce-306d-4f0b-aef7-af19955056fa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496531 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496553 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d9189ca-6d00-4596-980a-690341335c2a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gw7bx\" (UID: \"4d9189ca-6d00-4596-980a-690341335c2a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496579 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ntvl\" (UniqueName: \"kubernetes.io/projected/46dfa510-96af-4d45-9be7-5bd9a8588a61-kube-api-access-6ntvl\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.496605 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-metrics-tls\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499613 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c3149310-e8c4-4c36-8ff6-993af0481122-proxy-tls\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499636 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/97829faf-132d-43f9-8b6a-8b6e21c3f83e-signing-key\") pod \"service-ca-9c57cc56f-8j46g\" (UID: \"97829faf-132d-43f9-8b6a-8b6e21c3f83e\") " pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499702 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-certificates\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499749 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f25b1b4a-7676-4f7c-847a-ee29217bb19d-webhook-cert\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499765 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499781 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fgcx\" (UniqueName: \"kubernetes.io/projected/82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa-kube-api-access-6fgcx\") pod \"ingress-canary-6r6bw\" (UID: \"82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa\") " pod="openshift-ingress-canary/ingress-canary-6r6bw" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499808 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c3149310-e8c4-4c36-8ff6-993af0481122-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499824 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd9mq\" (UniqueName: \"kubernetes.io/projected/15efc57c-51cd-465b-9f45-ab52ae521e55-kube-api-access-qd9mq\") pod \"openshift-controller-manager-operator-756b6f6bc6-nt2xl\" (UID: \"15efc57c-51cd-465b-9f45-ab52ae521e55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499896 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvjnc\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-kube-api-access-lvjnc\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499913 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-serving-cert\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.499936 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9jvt\" (UniqueName: \"kubernetes.io/projected/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-kube-api-access-d9jvt\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.497082 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f25b1b4a-7676-4f7c-847a-ee29217bb19d-tmpfs\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.500117 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-csi-data-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.500133 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/32bd318f-c091-4913-889f-9097777ceb8e-srv-cert\") pod \"olm-operator-6b444d44fb-jplds\" (UID: \"32bd318f-c091-4913-889f-9097777ceb8e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.500151 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15efc57c-51cd-465b-9f45-ab52ae521e55-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nt2xl\" (UID: \"15efc57c-51cd-465b-9f45-ab52ae521e55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.508838 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c3149310-e8c4-4c36-8ff6-993af0481122-proxy-tls\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.511527 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c3149310-e8c4-4c36-8ff6-993af0481122-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.512048 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.512740 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-certificates\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.513853 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c3149310-e8c4-4c36-8ff6-993af0481122-images\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.519167 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.519789 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9pzvc"] Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.519938 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15efc57c-51cd-465b-9f45-ab52ae521e55-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-nt2xl\" (UID: \"15efc57c-51cd-465b-9f45-ab52ae521e55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.520168 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b915a9bf-4d3d-44ae-aa8f-243283528d77-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zxqjj\" (UID: \"b915a9bf-4d3d-44ae-aa8f-243283528d77\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.520718 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.522121 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/135afe15-d724-45ba-af8b-2a11862c594b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jqmsx\" (UID: \"135afe15-d724-45ba-af8b-2a11862c594b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526470 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/97829faf-132d-43f9-8b6a-8b6e21c3f83e-signing-cabundle\") pod \"service-ca-9c57cc56f-8j46g\" (UID: \"97829faf-132d-43f9-8b6a-8b6e21c3f83e\") " pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526554 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-bound-sa-token\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526588 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526606 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnmdq\" (UniqueName: \"kubernetes.io/projected/32bd318f-c091-4913-889f-9097777ceb8e-kube-api-access-qnmdq\") pod \"olm-operator-6b444d44fb-jplds\" (UID: \"32bd318f-c091-4913-889f-9097777ceb8e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526629 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526645 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpwpv\" (UniqueName: \"kubernetes.io/projected/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-kube-api-access-fpwpv\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526663 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aba2b97e-b9a0-4bc8-a480-84f64c2a0c55-proxy-tls\") pod \"machine-config-controller-84d6567774-lr2gg\" (UID: \"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526679 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4dbs\" (UniqueName: \"kubernetes.io/projected/c6523c2a-3325-4334-ba30-7323b1597aaf-kube-api-access-v4dbs\") pod \"catalog-operator-68c6474976-f7sbq\" (UID: \"c6523c2a-3325-4334-ba30-7323b1597aaf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526696 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-km2tl\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526722 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f25b1b4a-7676-4f7c-847a-ee29217bb19d-apiservice-cert\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526764 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-trusted-ca\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526844 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aba2b97e-b9a0-4bc8-a480-84f64c2a0c55-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lr2gg\" (UID: \"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526860 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8410b9e5-a83f-4bef-a99a-736c634d9e41-metrics-tls\") pod \"dns-default-9f9j7\" (UID: \"8410b9e5-a83f-4bef-a99a-736c634d9e41\") " pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526891 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v45dh\" (UniqueName: \"kubernetes.io/projected/399368ce-306d-4f0b-aef7-af19955056fa-kube-api-access-v45dh\") pod \"service-ca-operator-777779d784-sxhs2\" (UID: \"399368ce-306d-4f0b-aef7-af19955056fa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526907 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swk7s\" (UniqueName: \"kubernetes.io/projected/04ea9c16-02a9-4116-9a09-63a19de322bd-kube-api-access-swk7s\") pod \"machine-config-server-8rncq\" (UID: \"04ea9c16-02a9-4116-9a09-63a19de322bd\") " pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526946 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-tls\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526973 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.526992 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-encryption-config\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.527013 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bc07aea-e7b5-473c-957c-a5295edc2f4e-secret-volume\") pod \"collect-profiles-29432655-x4nw4\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.527038 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/569d9eb4-d037-4809-91a4-ee04d579cd21-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9q7zt\" (UID: \"569d9eb4-d037-4809-91a4-ee04d579cd21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.527056 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-897wc\" (UniqueName: \"kubernetes.io/projected/f25b1b4a-7676-4f7c-847a-ee29217bb19d-kube-api-access-897wc\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.527069 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8410b9e5-a83f-4bef-a99a-736c634d9e41-config-volume\") pod \"dns-default-9f9j7\" (UID: \"8410b9e5-a83f-4bef-a99a-736c634d9e41\") " pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.527101 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.527116 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/399368ce-306d-4f0b-aef7-af19955056fa-config\") pod \"service-ca-operator-777779d784-sxhs2\" (UID: \"399368ce-306d-4f0b-aef7-af19955056fa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.527134 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-trusted-ca\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.527179 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcpzl\" (UniqueName: \"kubernetes.io/projected/4d9189ca-6d00-4596-980a-690341335c2a-kube-api-access-xcpzl\") pod \"package-server-manager-789f6589d5-gw7bx\" (UID: \"4d9189ca-6d00-4596-980a-690341335c2a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.536021 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4d9189ca-6d00-4596-980a-690341335c2a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gw7bx\" (UID: \"4d9189ca-6d00-4596-980a-690341335c2a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.536937 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b915a9bf-4d3d-44ae-aa8f-243283528d77-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zxqjj\" (UID: \"b915a9bf-4d3d-44ae-aa8f-243283528d77\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.537426 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-tls\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.537957 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.538577 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c6523c2a-3325-4334-ba30-7323b1597aaf-srv-cert\") pod \"catalog-operator-68c6474976-f7sbq\" (UID: \"c6523c2a-3325-4334-ba30-7323b1597aaf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.539690 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f25b1b4a-7676-4f7c-847a-ee29217bb19d-webhook-cert\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.541053 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-trusted-ca\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.541337 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f25b1b4a-7676-4f7c-847a-ee29217bb19d-apiservice-cert\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.542288 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15efc57c-51cd-465b-9f45-ab52ae521e55-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-nt2xl\" (UID: \"15efc57c-51cd-465b-9f45-ab52ae521e55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.544721 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-metrics-tls\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.545440 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-trusted-ca\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: E1217 08:23:42.547546 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.033763899 +0000 UTC m=+158.578833901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.554784 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c6523c2a-3325-4334-ba30-7323b1597aaf-profile-collector-cert\") pod \"catalog-operator-68c6474976-f7sbq\" (UID: \"c6523c2a-3325-4334-ba30-7323b1597aaf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.557672 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.569950 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/569d9eb4-d037-4809-91a4-ee04d579cd21-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9q7zt\" (UID: \"569d9eb4-d037-4809-91a4-ee04d579cd21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.580158 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww9l2\" (UniqueName: \"kubernetes.io/projected/b6ac4609-63a9-400d-be84-cc6e4cb1a9cb-kube-api-access-ww9l2\") pod \"migrator-59844c95c7-c56mr\" (UID: \"b6ac4609-63a9-400d-be84-cc6e4cb1a9cb\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.627332 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r5fmr" event={"ID":"091f07a2-acd7-4d1e-b242-220c47716c25","Type":"ContainerStarted","Data":"839d61dae71732df4da2834c6c37a5a7fe85cf8f9c0c20f7df49b1a17cd9ffba"} Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.628626 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r5fmr" event={"ID":"091f07a2-acd7-4d1e-b242-220c47716c25","Type":"ContainerStarted","Data":"ad1a454a04ae1b5a15e729edd93b7e6d3977cdd137fccb8c1916210f9ea7b018"} Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.634705 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-r5fmr" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635434 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635613 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-audit-policies\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635634 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/399368ce-306d-4f0b-aef7-af19955056fa-serving-cert\") pod \"service-ca-operator-777779d784-sxhs2\" (UID: \"399368ce-306d-4f0b-aef7-af19955056fa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635654 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ntvl\" (UniqueName: \"kubernetes.io/projected/46dfa510-96af-4d45-9be7-5bd9a8588a61-kube-api-access-6ntvl\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635678 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/97829faf-132d-43f9-8b6a-8b6e21c3f83e-signing-key\") pod \"service-ca-9c57cc56f-8j46g\" (UID: \"97829faf-132d-43f9-8b6a-8b6e21c3f83e\") " pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635710 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fgcx\" (UniqueName: \"kubernetes.io/projected/82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa-kube-api-access-6fgcx\") pod \"ingress-canary-6r6bw\" (UID: \"82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa\") " pod="openshift-ingress-canary/ingress-canary-6r6bw" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635741 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-serving-cert\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635768 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9jvt\" (UniqueName: \"kubernetes.io/projected/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-kube-api-access-d9jvt\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635814 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-csi-data-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635837 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/32bd318f-c091-4913-889f-9097777ceb8e-srv-cert\") pod \"olm-operator-6b444d44fb-jplds\" (UID: \"32bd318f-c091-4913-889f-9097777ceb8e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635867 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/97829faf-132d-43f9-8b6a-8b6e21c3f83e-signing-cabundle\") pod \"service-ca-9c57cc56f-8j46g\" (UID: \"97829faf-132d-43f9-8b6a-8b6e21c3f83e\") " pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635908 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnmdq\" (UniqueName: \"kubernetes.io/projected/32bd318f-c091-4913-889f-9097777ceb8e-kube-api-access-qnmdq\") pod \"olm-operator-6b444d44fb-jplds\" (UID: \"32bd318f-c091-4913-889f-9097777ceb8e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635944 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.635976 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aba2b97e-b9a0-4bc8-a480-84f64c2a0c55-proxy-tls\") pod \"machine-config-controller-84d6567774-lr2gg\" (UID: \"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636014 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-km2tl\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636043 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aba2b97e-b9a0-4bc8-a480-84f64c2a0c55-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lr2gg\" (UID: \"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636072 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8410b9e5-a83f-4bef-a99a-736c634d9e41-metrics-tls\") pod \"dns-default-9f9j7\" (UID: \"8410b9e5-a83f-4bef-a99a-736c634d9e41\") " pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636094 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swk7s\" (UniqueName: \"kubernetes.io/projected/04ea9c16-02a9-4116-9a09-63a19de322bd-kube-api-access-swk7s\") pod \"machine-config-server-8rncq\" (UID: \"04ea9c16-02a9-4116-9a09-63a19de322bd\") " pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636118 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v45dh\" (UniqueName: \"kubernetes.io/projected/399368ce-306d-4f0b-aef7-af19955056fa-kube-api-access-v45dh\") pod \"service-ca-operator-777779d784-sxhs2\" (UID: \"399368ce-306d-4f0b-aef7-af19955056fa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636155 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-encryption-config\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636195 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bc07aea-e7b5-473c-957c-a5295edc2f4e-secret-volume\") pod \"collect-profiles-29432655-x4nw4\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636224 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8410b9e5-a83f-4bef-a99a-736c634d9e41-config-volume\") pod \"dns-default-9f9j7\" (UID: \"8410b9e5-a83f-4bef-a99a-736c634d9e41\") " pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636253 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/399368ce-306d-4f0b-aef7-af19955056fa-config\") pod \"service-ca-operator-777779d784-sxhs2\" (UID: \"399368ce-306d-4f0b-aef7-af19955056fa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636283 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa-cert\") pod \"ingress-canary-6r6bw\" (UID: \"82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa\") " pod="openshift-ingress-canary/ingress-canary-6r6bw" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636311 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-976x6\" (UniqueName: \"kubernetes.io/projected/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-kube-api-access-976x6\") pod \"marketplace-operator-79b997595-km2tl\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636332 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/04ea9c16-02a9-4116-9a09-63a19de322bd-certs\") pod \"machine-config-server-8rncq\" (UID: \"04ea9c16-02a9-4116-9a09-63a19de322bd\") " pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636354 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-registration-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636382 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d8c6\" (UniqueName: \"kubernetes.io/projected/aba2b97e-b9a0-4bc8-a480-84f64c2a0c55-kube-api-access-8d8c6\") pod \"machine-config-controller-84d6567774-lr2gg\" (UID: \"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636405 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-etcd-client\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636428 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thlgf\" (UniqueName: \"kubernetes.io/projected/97829faf-132d-43f9-8b6a-8b6e21c3f83e-kube-api-access-thlgf\") pod \"service-ca-9c57cc56f-8j46g\" (UID: \"97829faf-132d-43f9-8b6a-8b6e21c3f83e\") " pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636449 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-socket-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636473 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bc07aea-e7b5-473c-957c-a5295edc2f4e-config-volume\") pod \"collect-profiles-29432655-x4nw4\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636507 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9hvw\" (UniqueName: \"kubernetes.io/projected/8410b9e5-a83f-4bef-a99a-736c634d9e41-kube-api-access-s9hvw\") pod \"dns-default-9f9j7\" (UID: \"8410b9e5-a83f-4bef-a99a-736c634d9e41\") " pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636530 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-km2tl\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636552 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-k5zzh"] Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636558 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/32bd318f-c091-4913-889f-9097777ceb8e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jplds\" (UID: \"32bd318f-c091-4913-889f-9097777ceb8e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636639 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjknh\" (UniqueName: \"kubernetes.io/projected/9bc07aea-e7b5-473c-957c-a5295edc2f4e-kube-api-access-bjknh\") pod \"collect-profiles-29432655-x4nw4\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636663 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636685 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/04ea9c16-02a9-4116-9a09-63a19de322bd-node-bootstrap-token\") pod \"machine-config-server-8rncq\" (UID: \"04ea9c16-02a9-4116-9a09-63a19de322bd\") " pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636719 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-plugins-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636744 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-audit-dir\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636762 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-mountpoint-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.636899 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-mountpoint-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.637422 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: E1217 08:23:42.639910 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.13986099 +0000 UTC m=+158.684930962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.640811 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-audit-policies\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.648546 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-plugins-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.648646 4966 patch_prober.go:28] interesting pod/downloads-7954f5f757-r5fmr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.648682 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r5fmr" podUID="091f07a2-acd7-4d1e-b242-220c47716c25" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.649547 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-audit-dir\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.650090 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-csi-data-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.650237 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-registration-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.653768 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.661206 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/97829faf-132d-43f9-8b6a-8b6e21c3f83e-signing-cabundle\") pod \"service-ca-9c57cc56f-8j46g\" (UID: \"97829faf-132d-43f9-8b6a-8b6e21c3f83e\") " pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.655021 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-922bk" event={"ID":"c7bbf1fa-fe65-456b-97ff-c57e2786ad73","Type":"ContainerStarted","Data":"d42e3313755f8a89e8a85ff2ad0343b232a23ec8b3d0a032006d01c44ab30408"} Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.661744 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-serving-cert\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.662676 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bc07aea-e7b5-473c-957c-a5295edc2f4e-config-volume\") pod \"collect-profiles-29432655-x4nw4\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.662784 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/46dfa510-96af-4d45-9be7-5bd9a8588a61-socket-dir\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.665570 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-km2tl\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.665580 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.667157 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/aba2b97e-b9a0-4bc8-a480-84f64c2a0c55-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lr2gg\" (UID: \"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.668134 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" event={"ID":"e0296a9b-6991-4eb3-81d4-b7c710bc9acf","Type":"ContainerStarted","Data":"640ae038d883cd6e7c779cc7ff577ac1e3ecddb422758618d3c342c6701c5464"} Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.668280 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" event={"ID":"e0296a9b-6991-4eb3-81d4-b7c710bc9acf","Type":"ContainerStarted","Data":"9f44cf8f8f2a493dfb184ae8385de3f1273c185b8296f0edd113683bfe971e40"} Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.669468 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/399368ce-306d-4f0b-aef7-af19955056fa-config\") pod \"service-ca-operator-777779d784-sxhs2\" (UID: \"399368ce-306d-4f0b-aef7-af19955056fa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.670134 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/97829faf-132d-43f9-8b6a-8b6e21c3f83e-signing-key\") pod \"service-ca-9c57cc56f-8j46g\" (UID: \"97829faf-132d-43f9-8b6a-8b6e21c3f83e\") " pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.674974 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.675705 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/32bd318f-c091-4913-889f-9097777ceb8e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jplds\" (UID: \"32bd318f-c091-4913-889f-9097777ceb8e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.676677 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd9mq\" (UniqueName: \"kubernetes.io/projected/15efc57c-51cd-465b-9f45-ab52ae521e55-kube-api-access-qd9mq\") pod \"openshift-controller-manager-operator-756b6f6bc6-nt2xl\" (UID: \"15efc57c-51cd-465b-9f45-ab52ae521e55\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.677556 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8410b9e5-a83f-4bef-a99a-736c634d9e41-metrics-tls\") pod \"dns-default-9f9j7\" (UID: \"8410b9e5-a83f-4bef-a99a-736c634d9e41\") " pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.678517 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-etcd-client\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.679463 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8410b9e5-a83f-4bef-a99a-736c634d9e41-config-volume\") pod \"dns-default-9f9j7\" (UID: \"8410b9e5-a83f-4bef-a99a-736c634d9e41\") " pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.680084 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bc07aea-e7b5-473c-957c-a5295edc2f4e-secret-volume\") pod \"collect-profiles-29432655-x4nw4\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.680215 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-km2tl\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.680466 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b915a9bf-4d3d-44ae-aa8f-243283528d77-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zxqjj\" (UID: \"b915a9bf-4d3d-44ae-aa8f-243283528d77\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.681025 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmflq\" (UniqueName: \"kubernetes.io/projected/c3149310-e8c4-4c36-8ff6-993af0481122-kube-api-access-cmflq\") pod \"machine-config-operator-74547568cd-ktkmn\" (UID: \"c3149310-e8c4-4c36-8ff6-993af0481122\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.681559 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/aba2b97e-b9a0-4bc8-a480-84f64c2a0c55-proxy-tls\") pod \"machine-config-controller-84d6567774-lr2gg\" (UID: \"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.688612 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/04ea9c16-02a9-4116-9a09-63a19de322bd-node-bootstrap-token\") pod \"machine-config-server-8rncq\" (UID: \"04ea9c16-02a9-4116-9a09-63a19de322bd\") " pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.698319 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-b998v" event={"ID":"992684e8-e50b-44e7-b30c-7169f4877695","Type":"ContainerStarted","Data":"4cfcdc71daf95981cede6a1817eda7862ce53866d79817de2661fd4e9c998bca"} Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.698525 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/399368ce-306d-4f0b-aef7-af19955056fa-serving-cert\") pod \"service-ca-operator-777779d784-sxhs2\" (UID: \"399368ce-306d-4f0b-aef7-af19955056fa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.701770 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvjnc\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-kube-api-access-lvjnc\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.705981 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6db5f\" (UniqueName: \"kubernetes.io/projected/d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd-kube-api-access-6db5f\") pod \"ingress-operator-5b745b69d9-cjclj\" (UID: \"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.709111 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-encryption-config\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.712641 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa-cert\") pod \"ingress-canary-6r6bw\" (UID: \"82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa\") " pod="openshift-ingress-canary/ingress-canary-6r6bw" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.720466 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrqzd\" (UniqueName: \"kubernetes.io/projected/135afe15-d724-45ba-af8b-2a11862c594b-kube-api-access-mrqzd\") pod \"control-plane-machine-set-operator-78cbb6b69f-jqmsx\" (UID: \"135afe15-d724-45ba-af8b-2a11862c594b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.754457 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/04ea9c16-02a9-4116-9a09-63a19de322bd-certs\") pod \"machine-config-server-8rncq\" (UID: \"04ea9c16-02a9-4116-9a09-63a19de322bd\") " pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.754690 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/32bd318f-c091-4913-889f-9097777ceb8e-srv-cert\") pod \"olm-operator-6b444d44fb-jplds\" (UID: \"32bd318f-c091-4913-889f-9097777ceb8e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.764217 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: E1217 08:23:42.765628 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.265614658 +0000 UTC m=+158.810684600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.864179 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.864737 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-bound-sa-token\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.864921 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:42 crc kubenswrapper[4966]: W1217 08:23:42.864926 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ab4a948_e08e_4c50_a900_cfab1e8b7d4f.slice/crio-d485d204996b9dd93dc4cd966b2e64f591e86821019e2cd02ff0f3fdecce0816 WatchSource:0}: Error finding container d485d204996b9dd93dc4cd966b2e64f591e86821019e2cd02ff0f3fdecce0816: Status 404 returned error can't find the container with id d485d204996b9dd93dc4cd966b2e64f591e86821019e2cd02ff0f3fdecce0816 Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.873119 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpwpv\" (UniqueName: \"kubernetes.io/projected/12c4901f-3d9d-40c6-a224-e31a3f0d8b97-kube-api-access-fpwpv\") pod \"cluster-image-registry-operator-dc59b4c8b-j9jk4\" (UID: \"12c4901f-3d9d-40c6-a224-e31a3f0d8b97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:42 crc kubenswrapper[4966]: E1217 08:23:42.873265 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.373225332 +0000 UTC m=+158.918295274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.873960 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.876793 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4dbs\" (UniqueName: \"kubernetes.io/projected/c6523c2a-3325-4334-ba30-7323b1597aaf-kube-api-access-v4dbs\") pod \"catalog-operator-68c6474976-f7sbq\" (UID: \"c6523c2a-3325-4334-ba30-7323b1597aaf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.879501 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/569d9eb4-d037-4809-91a4-ee04d579cd21-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9q7zt\" (UID: \"569d9eb4-d037-4809-91a4-ee04d579cd21\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.879703 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.880286 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.881687 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.882260 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.882579 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.913426 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.942764 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:42 crc kubenswrapper[4966]: E1217 08:23:42.943210 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.443173162 +0000 UTC m=+158.988243104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.957095 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" event={"ID":"b053b6a3-d480-4dee-9c7f-146c7d084e06","Type":"ContainerStarted","Data":"f816cd6bb7518b200d1170410d2cde8ff9ad0e9fb8b6d50eac3f1ef9db10da0c"} Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.961264 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-897wc\" (UniqueName: \"kubernetes.io/projected/f25b1b4a-7676-4f7c-847a-ee29217bb19d-kube-api-access-897wc\") pod \"packageserver-d55dfcdfc-j476q\" (UID: \"f25b1b4a-7676-4f7c-847a-ee29217bb19d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.962081 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fgcx\" (UniqueName: \"kubernetes.io/projected/82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa-kube-api-access-6fgcx\") pod \"ingress-canary-6r6bw\" (UID: \"82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa\") " pod="openshift-ingress-canary/ingress-canary-6r6bw" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.962319 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" event={"ID":"1b60ed2c-77ac-4cbc-9b57-48ed79419073","Type":"ContainerStarted","Data":"4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527"} Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.962344 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" event={"ID":"1b60ed2c-77ac-4cbc-9b57-48ed79419073","Type":"ContainerStarted","Data":"41acd75f509ba5f4494933aef48f5fab95fc579e70b34fd723cc9abcbdb0333c"} Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.962749 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.978520 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:42 crc kubenswrapper[4966]: E1217 08:23:42.980136 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.480119555 +0000 UTC m=+159.025189497 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.994660 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6r6bw" Dec 17 08:23:42 crc kubenswrapper[4966]: I1217 08:23:42.995672 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcpzl\" (UniqueName: \"kubernetes.io/projected/4d9189ca-6d00-4596-980a-690341335c2a-kube-api-access-xcpzl\") pod \"package-server-manager-789f6589d5-gw7bx\" (UID: \"4d9189ca-6d00-4596-980a-690341335c2a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.027617 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9jvt\" (UniqueName: \"kubernetes.io/projected/2fd94046-2a2c-45c3-9922-0962c7f7b3eb-kube-api-access-d9jvt\") pod \"apiserver-7bbb656c7d-vtwwx\" (UID: \"2fd94046-2a2c-45c3-9922-0962c7f7b3eb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.030012 4966 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-gzzdt container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.030047 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" podUID="1b60ed2c-77ac-4cbc-9b57-48ed79419073" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.030717 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ntvl\" (UniqueName: \"kubernetes.io/projected/46dfa510-96af-4d45-9be7-5bd9a8588a61-kube-api-access-6ntvl\") pod \"csi-hostpathplugin-lcqdh\" (UID: \"46dfa510-96af-4d45-9be7-5bd9a8588a61\") " pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.033388 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxfxg"] Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.034292 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" event={"ID":"41dbc543-cace-4c8f-8db3-e07b984c0919","Type":"ContainerStarted","Data":"71a67e5f2925413a5df9fe692efd5a759414ef5c38425da0513628cddbca3f63"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.036073 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjknh\" (UniqueName: \"kubernetes.io/projected/9bc07aea-e7b5-473c-957c-a5295edc2f4e-kube-api-access-bjknh\") pod \"collect-profiles-29432655-x4nw4\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.063859 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" event={"ID":"7a83b3e0-2092-491a-8844-1aec20fb2752","Type":"ContainerStarted","Data":"4b50fc62bfbcb960297bfd1433e9efa8ad4100acd04a9291cd458d6a06fc26de"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.064195 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" event={"ID":"7a83b3e0-2092-491a-8844-1aec20fb2752","Type":"ContainerStarted","Data":"110ffb4ac084c2d6c96213776007877812c7874c0b9a672fe1848bfa0904d7b6"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.066389 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" event={"ID":"bff81677-e3e6-4e74-81ba-d89156398d45","Type":"ContainerStarted","Data":"0bf7b5316bad596a9ad4e77b8a98b51ec82c0c93e546b606918e5cf0bfe3da3d"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.066419 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" event={"ID":"bff81677-e3e6-4e74-81ba-d89156398d45","Type":"ContainerStarted","Data":"d4804fb38d86da0aa3c4b4db3626158f1985617cc9504b3d7fb2e3fa9adb2ef6"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.068168 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" event={"ID":"a5cc70d0-a401-46b1-b102-a397fa0183c7","Type":"ContainerStarted","Data":"de35c697a2b535704b60028cb0e08c1828de682730886a22b8f5fe6d1f7e3d30"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.068196 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" event={"ID":"a5cc70d0-a401-46b1-b102-a397fa0183c7","Type":"ContainerStarted","Data":"8a9ed51fd4a84f4db79e61f9770fa9e33442ae40b1cfa56b5a5bd1129407d6bf"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.069430 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" event={"ID":"50255104-8dc1-485a-b18b-b2f86d9150b4","Type":"ContainerStarted","Data":"ebe9264125d40f2409386f5e993951debab87716b59ecc88a75e6e7b05d46fdc"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.071496 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnmdq\" (UniqueName: \"kubernetes.io/projected/32bd318f-c091-4913-889f-9097777ceb8e-kube-api-access-qnmdq\") pod \"olm-operator-6b444d44fb-jplds\" (UID: \"32bd318f-c091-4913-889f-9097777ceb8e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.079189 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v45dh\" (UniqueName: \"kubernetes.io/projected/399368ce-306d-4f0b-aef7-af19955056fa-kube-api-access-v45dh\") pod \"service-ca-operator-777779d784-sxhs2\" (UID: \"399368ce-306d-4f0b-aef7-af19955056fa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.079834 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swk7s\" (UniqueName: \"kubernetes.io/projected/04ea9c16-02a9-4116-9a09-63a19de322bd-kube-api-access-swk7s\") pod \"machine-config-server-8rncq\" (UID: \"04ea9c16-02a9-4116-9a09-63a19de322bd\") " pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.080157 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d8c6\" (UniqueName: \"kubernetes.io/projected/aba2b97e-b9a0-4bc8-a480-84f64c2a0c55-kube-api-access-8d8c6\") pod \"machine-config-controller-84d6567774-lr2gg\" (UID: \"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.080214 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9hvw\" (UniqueName: \"kubernetes.io/projected/8410b9e5-a83f-4bef-a99a-736c634d9e41-kube-api-access-s9hvw\") pod \"dns-default-9f9j7\" (UID: \"8410b9e5-a83f-4bef-a99a-736c634d9e41\") " pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.080817 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.101789 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thlgf\" (UniqueName: \"kubernetes.io/projected/97829faf-132d-43f9-8b6a-8b6e21c3f83e-kube-api-access-thlgf\") pod \"service-ca-9c57cc56f-8j46g\" (UID: \"97829faf-132d-43f9-8b6a-8b6e21c3f83e\") " pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.102055 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-976x6\" (UniqueName: \"kubernetes.io/projected/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-kube-api-access-976x6\") pod \"marketplace-operator-79b997595-km2tl\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.102383 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.107528 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.607503848 +0000 UTC m=+159.152573790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.137833 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-587gm" event={"ID":"8b1849b5-04d6-4550-bd09-0f69c712b283","Type":"ContainerStarted","Data":"a08f90b617b481adac64d1e2350620919db213d26efa142fb08fa950021fcb68"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.157636 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" event={"ID":"906afeba-a01c-4762-921a-35f608da53b2","Type":"ContainerStarted","Data":"e8b9599dc420eb51a7ec4054038b8ed017e5dbd5ebb786563f96ab42bb58e611"} Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.211297 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.213137 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.216514 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.217844 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.717822098 +0000 UTC m=+159.262892040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.239027 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.239452 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.240725 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.250981 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.270406 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.270892 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.271127 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.278675 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.305727 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.313131 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8rncq" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.318291 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.318580 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.818567771 +0000 UTC m=+159.363637703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.419781 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.420270 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:43.920249221 +0000 UTC m=+159.465319163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.428612 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md"] Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.482826 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-clm56"] Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.500289 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd"] Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.521522 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.522097 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.022085575 +0000 UTC m=+159.567155517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.624475 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.624966 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.124943878 +0000 UTC m=+159.670013820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.625190 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.625483 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.125475812 +0000 UTC m=+159.670545754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.637234 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" podStartSLOduration=140.637213938 podStartE2EDuration="2m20.637213938s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:43.636399225 +0000 UTC m=+159.181469167" watchObservedRunningTime="2025-12-17 08:23:43.637213938 +0000 UTC m=+159.182283880" Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.730553 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.730921 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.230905896 +0000 UTC m=+159.775975838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.831596 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.831935 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.331920497 +0000 UTC m=+159.876990439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.936559 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:43 crc kubenswrapper[4966]: E1217 08:23:43.936911 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.436896788 +0000 UTC m=+159.981966730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:43 crc kubenswrapper[4966]: I1217 08:23:43.963082 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" podStartSLOduration=140.963063113 podStartE2EDuration="2m20.963063113s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:43.962971841 +0000 UTC m=+159.508041793" watchObservedRunningTime="2025-12-17 08:23:43.963063113 +0000 UTC m=+159.508133045" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.023641 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr"] Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.060614 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.061507 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.561489383 +0000 UTC m=+160.106559325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.076539 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx"] Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.125664 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn"] Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.157991 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj"] Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.162466 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.162813 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.662798172 +0000 UTC m=+160.207868104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.242990 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-r5fmr" podStartSLOduration=141.242971595 podStartE2EDuration="2m21.242971595s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:44.23520748 +0000 UTC m=+159.780277422" watchObservedRunningTime="2025-12-17 08:23:44.242971595 +0000 UTC m=+159.788041527" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.243783 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt"] Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.270964 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.271286 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.771272961 +0000 UTC m=+160.316342903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: W1217 08:23:44.281549 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod135afe15_d724_45ba_af8b_2a11862c594b.slice/crio-7189a4eb0deeda4b9d648849dd13dde026943c7fb3aeb7610d3e95cc54dfbcb2 WatchSource:0}: Error finding container 7189a4eb0deeda4b9d648849dd13dde026943c7fb3aeb7610d3e95cc54dfbcb2: Status 404 returned error can't find the container with id 7189a4eb0deeda4b9d648849dd13dde026943c7fb3aeb7610d3e95cc54dfbcb2 Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.372174 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.372451 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.872437096 +0000 UTC m=+160.417507038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: W1217 08:23:44.426983 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3149310_e8c4_4c36_8ff6_993af0481122.slice/crio-0ee9eb42060e4cbab47bba479c9359de6654d17654c1a54616b6f15c0b59c333 WatchSource:0}: Error finding container 0ee9eb42060e4cbab47bba479c9359de6654d17654c1a54616b6f15c0b59c333: Status 404 returned error can't find the container with id 0ee9eb42060e4cbab47bba479c9359de6654d17654c1a54616b6f15c0b59c333 Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.427722 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" event={"ID":"bff81677-e3e6-4e74-81ba-d89156398d45","Type":"ContainerStarted","Data":"46823f5397c9bf4c4de0a3add27d5310abc02f19cf7a1d321b20e173be96f115"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.471512 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" event={"ID":"50255104-8dc1-485a-b18b-b2f86d9150b4","Type":"ContainerStarted","Data":"7a8812f2a91c7f457a17e4ac622a125c2481ff696be7cbb5ec90403c95da298a"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.476642 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.477084 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:44.977070277 +0000 UTC m=+160.522140229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.503089 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-phjzd" podStartSLOduration=141.503067218 podStartE2EDuration="2m21.503067218s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:44.460441376 +0000 UTC m=+160.005511318" watchObservedRunningTime="2025-12-17 08:23:44.503067218 +0000 UTC m=+160.048137160" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.504803 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-b998v" event={"ID":"992684e8-e50b-44e7-b30c-7169f4877695","Type":"ContainerStarted","Data":"1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.521756 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" event={"ID":"5ab4a948-e08e-4c50-a900-cfab1e8b7d4f","Type":"ContainerStarted","Data":"d485d204996b9dd93dc4cd966b2e64f591e86821019e2cd02ff0f3fdecce0816"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.540431 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-587gm" event={"ID":"8b1849b5-04d6-4550-bd09-0f69c712b283","Type":"ContainerStarted","Data":"fcfce466189142efc2255d3ca1494af9d7d5dcb5b7a61570b297c9512437d6c4"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.541970 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" event={"ID":"88d927f5-13ac-470d-8544-9a4aa1dc55ff","Type":"ContainerStarted","Data":"fd0a83e344151c6c605a31b0b9b2491493561e752849e4c6623568bcf8d1de23"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.546179 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" event={"ID":"fadc152c-c5d3-4cec-b2bb-c35126687c76","Type":"ContainerStarted","Data":"2d2bf433a1c539110520e9803997ff797cec99e893e24ecd8fe9f34b4c1a6644"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.585345 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.586468 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.08645029 +0000 UTC m=+160.631520232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.596308 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" event={"ID":"a5cc70d0-a401-46b1-b102-a397fa0183c7","Type":"ContainerStarted","Data":"787fba211ec2ff66a0c8b9d8957acf027dbd8440a52dddf39b094f1ad390e822"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.662987 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" event={"ID":"dae6e8c2-ac2d-479b-83c1-c10aff34a548","Type":"ContainerStarted","Data":"a20a2e0b3b7cf3c0e7cbac29ac71dd319feeaa8f045e9fbcdcc4de1ae85dd148"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.686436 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.687567 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.187557354 +0000 UTC m=+160.732627286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.689053 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" event={"ID":"eb93a1c3-72e1-4fe1-96c3-742263dced3e","Type":"ContainerStarted","Data":"69aba48fe56561e4168120f5c0716bda854c6ee1edb78f6ee13de7506693bf34"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.724858 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl"] Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.760573 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq"] Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.774522 4966 generic.go:334] "Generic (PLEG): container finished" podID="7a83b3e0-2092-491a-8844-1aec20fb2752" containerID="4b50fc62bfbcb960297bfd1433e9efa8ad4100acd04a9291cd458d6a06fc26de" exitCode=0 Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.774671 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" event={"ID":"7a83b3e0-2092-491a-8844-1aec20fb2752","Type":"ContainerDied","Data":"4b50fc62bfbcb960297bfd1433e9efa8ad4100acd04a9291cd458d6a06fc26de"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.777889 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" event={"ID":"906afeba-a01c-4762-921a-35f608da53b2","Type":"ContainerStarted","Data":"287f8b9dd1b3ec22a30333f758da53dff0d7fc81dd7dc5f0fc75b9c51d8ca13d"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.779006 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" event={"ID":"021d4c31-6de9-4e76-ac80-0f345b3d7f81","Type":"ContainerStarted","Data":"72522b04e697435317f15de15d714dcd91f69c45b90c8733913d327b93bf61a6"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.787191 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.787337 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.28731465 +0000 UTC m=+160.832384592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.787453 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.788096 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.288088162 +0000 UTC m=+160.833158104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.819944 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" event={"ID":"b053b6a3-d480-4dee-9c7f-146c7d084e06","Type":"ContainerStarted","Data":"7ccb910183f9f132d433aae82ef9200f8c6f29d3d94a71e78c8fb835000dfcb6"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.821001 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.826948 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-b998v" podStartSLOduration=141.826914018 podStartE2EDuration="2m21.826914018s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:44.826180238 +0000 UTC m=+160.371250180" watchObservedRunningTime="2025-12-17 08:23:44.826914018 +0000 UTC m=+160.371983960" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.828288 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" event={"ID":"41dbc543-cace-4c8f-8db3-e07b984c0919","Type":"ContainerStarted","Data":"dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.828329 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.831993 4966 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-v6b7k container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.831997 4966 patch_prober.go:28] interesting pod/console-operator-58897d9998-f4bsv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.832031 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" podUID="41dbc543-cace-4c8f-8db3-e07b984c0919" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.832039 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" podUID="b053b6a3-d480-4dee-9c7f-146c7d084e06" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.834402 4966 generic.go:334] "Generic (PLEG): container finished" podID="c7bbf1fa-fe65-456b-97ff-c57e2786ad73" containerID="71900025e2c7ec74ec195e4ac268c2eb2e92ef6bee0942c88510d7ee690da812" exitCode=0 Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.837797 4966 patch_prober.go:28] interesting pod/downloads-7954f5f757-r5fmr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.837929 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r5fmr" podUID="091f07a2-acd7-4d1e-b242-220c47716c25" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.838966 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-bnjl5" podStartSLOduration=141.838936222 podStartE2EDuration="2m21.838936222s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:44.838131679 +0000 UTC m=+160.383201621" watchObservedRunningTime="2025-12-17 08:23:44.838936222 +0000 UTC m=+160.384006164" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.869049 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-587gm" podStartSLOduration=141.869026516 podStartE2EDuration="2m21.869026516s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:44.868206603 +0000 UTC m=+160.413276565" watchObservedRunningTime="2025-12-17 08:23:44.869026516 +0000 UTC m=+160.414096458" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.871488 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-j6vwq" podStartSLOduration=141.871479904 podStartE2EDuration="2m21.871479904s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:44.85330665 +0000 UTC m=+160.398376592" watchObservedRunningTime="2025-12-17 08:23:44.871479904 +0000 UTC m=+160.416549846" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.890922 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.891820 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.391805968 +0000 UTC m=+160.936875910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.893288 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" podStartSLOduration=141.893269159 podStartE2EDuration="2m21.893269159s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:44.892056455 +0000 UTC m=+160.437126407" watchObservedRunningTime="2025-12-17 08:23:44.893269159 +0000 UTC m=+160.438339101" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.908072 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" podStartSLOduration=141.908055288 podStartE2EDuration="2m21.908055288s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:44.908019177 +0000 UTC m=+160.453089119" watchObservedRunningTime="2025-12-17 08:23:44.908055288 +0000 UTC m=+160.453125250" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.951465 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.951512 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-922bk" event={"ID":"c7bbf1fa-fe65-456b-97ff-c57e2786ad73","Type":"ContainerDied","Data":"71900025e2c7ec74ec195e4ac268c2eb2e92ef6bee0942c88510d7ee690da812"} Dec 17 08:23:44 crc kubenswrapper[4966]: I1217 08:23:44.993677 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:44 crc kubenswrapper[4966]: E1217 08:23:44.994113 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.494091814 +0000 UTC m=+161.039161756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.094760 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:45 crc kubenswrapper[4966]: E1217 08:23:45.096662 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.596644848 +0000 UTC m=+161.141714790 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:45 crc kubenswrapper[4966]: W1217 08:23:45.138433 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6523c2a_3325_4334_ba30_7323b1597aaf.slice/crio-12470f72846b23b78dd302f56636d1ecfab4a98b2bc6128af627de456fb4b303 WatchSource:0}: Error finding container 12470f72846b23b78dd302f56636d1ecfab4a98b2bc6128af627de456fb4b303: Status 404 returned error can't find the container with id 12470f72846b23b78dd302f56636d1ecfab4a98b2bc6128af627de456fb4b303 Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.206003 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:45 crc kubenswrapper[4966]: E1217 08:23:45.206286 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.706273658 +0000 UTC m=+161.251343600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.281160 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.312108 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.329624 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:45 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:45 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:45 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.329687 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:45 crc kubenswrapper[4966]: E1217 08:23:45.331028 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.830999046 +0000 UTC m=+161.376068988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.387025 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4"] Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.432165 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:45 crc kubenswrapper[4966]: E1217 08:23:45.432467 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:45.93245448 +0000 UTC m=+161.477524412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.525183 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6r6bw"] Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.540495 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:45 crc kubenswrapper[4966]: E1217 08:23:45.540772 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:46.040747653 +0000 UTC m=+161.585817595 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.616721 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-km2tl"] Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.651630 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.651690 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:45 crc kubenswrapper[4966]: E1217 08:23:45.656036 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:46.15602034 +0000 UTC m=+161.701090282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.681558 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf617b8-bd09-49de-b188-9c86d5fafd57-metrics-certs\") pod \"network-metrics-daemon-nqmws\" (UID: \"6cf617b8-bd09-49de-b188-9c86d5fafd57\") " pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.705240 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx"] Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.718310 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj"] Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.725971 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8j46g"] Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.760983 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lcqdh"] Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.761345 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:45 crc kubenswrapper[4966]: E1217 08:23:45.761829 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:46.261815003 +0000 UTC m=+161.806884945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.809675 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2"] Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.862330 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx"] Dec 17 08:23:45 crc kubenswrapper[4966]: I1217 08:23:45.862957 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:45 crc kubenswrapper[4966]: E1217 08:23:45.863323 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:46.363311688 +0000 UTC m=+161.908381630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:45.876368 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg"] Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:45.930328 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" event={"ID":"7a83b3e0-2092-491a-8844-1aec20fb2752","Type":"ContainerStarted","Data":"e4c8d812e298bc3a670a09ef88f8f156468b7955484c2c4d62b56d60a9ba2e81"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:45.931632 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:45.964137 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:45.964519 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:46.464505464 +0000 UTC m=+162.009575406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:45.967697 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nqmws" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:45.971964 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9f9j7"] Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:45.992793 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" podStartSLOduration=142.992775357 podStartE2EDuration="2m22.992775357s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:45.984496088 +0000 UTC m=+161.529566030" watchObservedRunningTime="2025-12-17 08:23:45.992775357 +0000 UTC m=+161.537845299" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:45.994530 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4"] Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.004995 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q"] Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.043631 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds"] Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.060838 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" event={"ID":"906afeba-a01c-4762-921a-35f608da53b2","Type":"ContainerStarted","Data":"fd2186413743636be9610f32d6cd0e0a3456b2057875b14110477ed978dd51cf"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.068998 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:46.069313 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:46.56929581 +0000 UTC m=+162.114365752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.134140 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-75n4w" podStartSLOduration=143.134118397 podStartE2EDuration="2m23.134118397s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:46.133275713 +0000 UTC m=+161.678345665" watchObservedRunningTime="2025-12-17 08:23:46.134118397 +0000 UTC m=+161.679188339" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.187107 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:46.188021 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:46.688006432 +0000 UTC m=+162.233076374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.216951 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" event={"ID":"c3149310-e8c4-4c36-8ff6-993af0481122","Type":"ContainerStarted","Data":"1c2bb1018a47945167135c7b2888667e2ffd70d55c34668a8bef9b9a562a7149"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.216987 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" event={"ID":"c3149310-e8c4-4c36-8ff6-993af0481122","Type":"ContainerStarted","Data":"0ee9eb42060e4cbab47bba479c9359de6654d17654c1a54616b6f15c0b59c333"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.301362 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" event={"ID":"021d4c31-6de9-4e76-ac80-0f345b3d7f81","Type":"ContainerStarted","Data":"bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.304667 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.305693 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:46 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:46 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:46 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.305742 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.306542 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:46.311641 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:46.81161759 +0000 UTC m=+162.356687532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.344697 4966 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nxfxg container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.37:6443/healthz\": dial tcp 10.217.0.37:6443: connect: connection refused" start-of-body= Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.344759 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" podUID="021d4c31-6de9-4e76-ac80-0f345b3d7f81" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.37:6443/healthz\": dial tcp 10.217.0.37:6443: connect: connection refused" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.353890 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" podStartSLOduration=143.353860011 podStartE2EDuration="2m23.353860011s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:46.212632925 +0000 UTC m=+161.757702877" watchObservedRunningTime="2025-12-17 08:23:46.353860011 +0000 UTC m=+161.898929953" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.369908 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" event={"ID":"5ab4a948-e08e-4c50-a900-cfab1e8b7d4f","Type":"ContainerStarted","Data":"1e5af3c428620002998ffb51e4646b65a149fec77b79570423b0078bef650878"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.406816 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" event={"ID":"12c4901f-3d9d-40c6-a224-e31a3f0d8b97","Type":"ContainerStarted","Data":"04b12e3ef36b7f00606daf0b66d127ae709aac98c94377b6a19ebd938e41367b"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.407440 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6r6bw" event={"ID":"82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa","Type":"ContainerStarted","Data":"d696d9cf9d08eabf35f08073314b56c1ae6e9effef94fde2652453a404e3ef61"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.408318 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr" event={"ID":"b6ac4609-63a9-400d-be84-cc6e4cb1a9cb","Type":"ContainerStarted","Data":"c907cab2eebb6e36b785bfbf40bed2f312afbdeab5cdd167e0fb9dec9b214394"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.408345 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr" event={"ID":"b6ac4609-63a9-400d-be84-cc6e4cb1a9cb","Type":"ContainerStarted","Data":"a05143db54702c0d42dd5dad3d47286c395ca4ffec4c4a9e76840f22988a0a2b"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.409243 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" event={"ID":"135afe15-d724-45ba-af8b-2a11862c594b","Type":"ContainerStarted","Data":"5cbb1d82c39627832902f6b31506dbeeab4d010fb6d72f40b2ed2ce56484c413"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.409270 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" event={"ID":"135afe15-d724-45ba-af8b-2a11862c594b","Type":"ContainerStarted","Data":"7189a4eb0deeda4b9d648849dd13dde026943c7fb3aeb7610d3e95cc54dfbcb2"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.410550 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8rncq" event={"ID":"04ea9c16-02a9-4116-9a09-63a19de322bd","Type":"ContainerStarted","Data":"1ad1f964ba1622158ced96e0969c8672fd483e8647979aabc71091837a5f5862"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.411351 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" event={"ID":"15efc57c-51cd-465b-9f45-ab52ae521e55","Type":"ContainerStarted","Data":"628fabd24c5a1561d7b94a7c13232611e624e800931548fd0421c280d5351fa3"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.427691 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:46.428985 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:46.928968774 +0000 UTC m=+162.474038716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.458132 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" podStartSLOduration=143.458117221 podStartE2EDuration="2m23.458117221s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:46.368186348 +0000 UTC m=+161.913256290" watchObservedRunningTime="2025-12-17 08:23:46.458117221 +0000 UTC m=+162.003187173" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.509441 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" event={"ID":"dae6e8c2-ac2d-479b-83c1-c10aff34a548","Type":"ContainerStarted","Data":"d9513b6677eb297979edea29b96468660b50362d3baf6dcf841807cb3b50ad35"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.525526 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" podStartSLOduration=143.525507261 podStartE2EDuration="2m23.525507261s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:46.459797878 +0000 UTC m=+162.004867810" watchObservedRunningTime="2025-12-17 08:23:46.525507261 +0000 UTC m=+162.070577203" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.531290 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" event={"ID":"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd","Type":"ContainerStarted","Data":"592fba0c6f15e2e61684a5cd595f9b820597215de2b7cf06810fe4fa3e27aeb1"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.531336 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" event={"ID":"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd","Type":"ContainerStarted","Data":"2c426a4135df4b2edfc933c2e4300e8b4af705a5de00ff949c679253b0e0ef09"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.534718 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:46.535164 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:47.035152708 +0000 UTC m=+162.580222650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.572557 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" event={"ID":"eb93a1c3-72e1-4fe1-96c3-742263dced3e","Type":"ContainerStarted","Data":"78965cd1612efd67cf1a9cd3fd0997a864c61a8fa0729dfb3f61780375c6312f"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.581110 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" event={"ID":"c6523c2a-3325-4334-ba30-7323b1597aaf","Type":"ContainerStarted","Data":"12470f72846b23b78dd302f56636d1ecfab4a98b2bc6128af627de456fb4b303"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.582859 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" event={"ID":"569d9eb4-d037-4809-91a4-ee04d579cd21","Type":"ContainerStarted","Data":"63b9eab62c5f7bf5d08f8da0d70f8199ba527be2fde853ae2a8875b3e3af5b44"} Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.585773 4966 patch_prober.go:28] interesting pod/downloads-7954f5f757-r5fmr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.585799 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r5fmr" podUID="091f07a2-acd7-4d1e-b242-220c47716c25" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.595214 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.607960 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jqmsx" podStartSLOduration=143.607944126 podStartE2EDuration="2m23.607944126s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:46.531156368 +0000 UTC m=+162.076226310" watchObservedRunningTime="2025-12-17 08:23:46.607944126 +0000 UTC m=+162.153014078" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.639380 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:46.641132 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:47.141105936 +0000 UTC m=+162.686175878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.693245 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" podStartSLOduration=143.693228881 podStartE2EDuration="2m23.693228881s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:46.608689587 +0000 UTC m=+162.153759539" watchObservedRunningTime="2025-12-17 08:23:46.693228881 +0000 UTC m=+162.238298823" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.693971 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-9pzvc" podStartSLOduration=143.693964582 podStartE2EDuration="2m23.693964582s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:46.691848493 +0000 UTC m=+162.236918435" watchObservedRunningTime="2025-12-17 08:23:46.693964582 +0000 UTC m=+162.239034524" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.741936 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:46.742259 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:47.24222692 +0000 UTC m=+162.787296862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.807212 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.807256 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.843638 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:46.844023 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:47.344007873 +0000 UTC m=+162.889077815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.946130 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:46 crc kubenswrapper[4966]: E1217 08:23:46.947559 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:47.447545994 +0000 UTC m=+162.992615936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:46 crc kubenswrapper[4966]: I1217 08:23:46.967233 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.089990 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.090331 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:47.590314293 +0000 UTC m=+163.135384235 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.197642 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.198230 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:47.698214955 +0000 UTC m=+163.243284897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.293558 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:47 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:47 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:47 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.293821 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.302431 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.302809 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:47.802794275 +0000 UTC m=+163.347864217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.416604 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.416933 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:47.916921789 +0000 UTC m=+163.461991731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.521183 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.521324 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.021300334 +0000 UTC m=+163.566370266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.521427 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.521747 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.021736106 +0000 UTC m=+163.566806048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.632169 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.632504 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.132489257 +0000 UTC m=+163.677559199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.632848 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.633207 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.133199537 +0000 UTC m=+163.678269479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.636915 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-knt89" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.663731 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" event={"ID":"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c","Type":"ContainerStarted","Data":"9b7e44bfc85d5c569d47566faf974889bc46e58a3d5001bac59873c10aef7ac8"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.666518 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" event={"ID":"eb93a1c3-72e1-4fe1-96c3-742263dced3e","Type":"ContainerStarted","Data":"270f26a0124a8ddf88bb4ae194989577dc15fab233c184b3a56759b27e561e5e"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.670163 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8gpfd" event={"ID":"fadc152c-c5d3-4cec-b2bb-c35126687c76","Type":"ContainerStarted","Data":"85e2ae3ed99d4f91e71e35ea3787d4c78ad7ab123f030d3b065e68b0282c822f"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.703812 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8rncq" event={"ID":"04ea9c16-02a9-4116-9a09-63a19de322bd","Type":"ContainerStarted","Data":"c7cf09e96b61e715c25610e817bffcbe403d6a58977aa2b3f90218798dce6984"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.719399 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9f9j7" event={"ID":"8410b9e5-a83f-4bef-a99a-736c634d9e41","Type":"ContainerStarted","Data":"495f0069398260d7ab8a7cb2efd655f7120d1eef485ee1a6c0b58fc54acbe446"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.726523 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" event={"ID":"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55","Type":"ContainerStarted","Data":"c6d8bf273cfb3891d6a8f9bf604d5e0e4be9c46b1ba859c4e09eced4d1db4a19"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.726735 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" event={"ID":"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55","Type":"ContainerStarted","Data":"d1c366ebf175b249deedaf6eed1a3ed86372b2046b71ed46f44c45015a1a1c78"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.727758 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" event={"ID":"569d9eb4-d037-4809-91a4-ee04d579cd21","Type":"ContainerStarted","Data":"1f0b9b5b73a3db9e6f4e50e90e205a6c253d14b6cd8070762df27f572d410813"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.733433 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.733968 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.233954111 +0000 UTC m=+163.779024043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.748618 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" event={"ID":"32bd318f-c091-4913-889f-9097777ceb8e","Type":"ContainerStarted","Data":"35a12b3c771ffb4dcf8cf3108d209d84796268a39bfe140c4dd56896e4e1c44b"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.748915 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" event={"ID":"32bd318f-c091-4913-889f-9097777ceb8e","Type":"ContainerStarted","Data":"00855b22df39c36ab6d0b4f20b9a16b7730e05505584ffecd10dafa2956ed1ee"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.750825 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.750933 4966 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-jplds container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.750981 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" podUID="32bd318f-c091-4913-889f-9097777ceb8e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.771671 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-clm56" podStartSLOduration=144.771651367 podStartE2EDuration="2m24.771651367s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:47.73822793 +0000 UTC m=+163.283297872" watchObservedRunningTime="2025-12-17 08:23:47.771651367 +0000 UTC m=+163.316721309" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.772630 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8rncq" podStartSLOduration=7.772624083 podStartE2EDuration="7.772624083s" podCreationTimestamp="2025-12-17 08:23:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:47.767304865 +0000 UTC m=+163.312374807" watchObservedRunningTime="2025-12-17 08:23:47.772624083 +0000 UTC m=+163.317694025" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.788521 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr" event={"ID":"b6ac4609-63a9-400d-be84-cc6e4cb1a9cb","Type":"ContainerStarted","Data":"a93b7e60e8d3ffc41efbf1b07283eb2c79e25b10f99cc5f32bf6ba91faf5d039"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.814241 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nqmws"] Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.827233 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" event={"ID":"5ab4a948-e08e-4c50-a900-cfab1e8b7d4f","Type":"ContainerStarted","Data":"31466d178e80e1fa0111c542b8370fc6e2f90d2a85dfcc416dde24bab4769d79"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.838776 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9q7zt" podStartSLOduration=144.838755278 podStartE2EDuration="2m24.838755278s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:47.838427168 +0000 UTC m=+163.383497120" watchObservedRunningTime="2025-12-17 08:23:47.838755278 +0000 UTC m=+163.383825220" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.839719 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.841045 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.3410299 +0000 UTC m=+163.886099832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.852186 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" event={"ID":"97829faf-132d-43f9-8b6a-8b6e21c3f83e","Type":"ContainerStarted","Data":"4a9bae5dc08bd14afccd6a0e8c882b55027dd11c11765538dfb8c485e8127a14"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.853117 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" event={"ID":"4d9189ca-6d00-4596-980a-690341335c2a","Type":"ContainerStarted","Data":"4ae0bfc5e85f98caac34f7ef4bbc428a187f2093cb6cbd0c2691e35288520227"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.854139 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-nt2xl" event={"ID":"15efc57c-51cd-465b-9f45-ab52ae521e55","Type":"ContainerStarted","Data":"1877b8316e29f91c2eefa20d8a4d9ef27383a20144167118af93e761ae19853c"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.855416 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" event={"ID":"46dfa510-96af-4d45-9be7-5bd9a8588a61","Type":"ContainerStarted","Data":"19dd2e734d5fe527c91a95b7e72f057c37d65e9c65140fede623fb12c360535b"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.856519 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" event={"ID":"d0dd07cd-b519-4ee3-adc4-f42c4b3b7cfd","Type":"ContainerStarted","Data":"a2351a2ece0724d1c45cd830c3575e4b200a6f6e2cf26bfcf52c61d033791546"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.878057 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" event={"ID":"9bc07aea-e7b5-473c-957c-a5295edc2f4e","Type":"ContainerStarted","Data":"c061623f0d05889623a8db49b4e7330bd34d798f9ff934112212fe7035f5f58b"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.880835 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-k5zzh" podStartSLOduration=144.880817634 podStartE2EDuration="2m24.880817634s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:47.878490109 +0000 UTC m=+163.423560061" watchObservedRunningTime="2025-12-17 08:23:47.880817634 +0000 UTC m=+163.425887576" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.889042 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" event={"ID":"f25b1b4a-7676-4f7c-847a-ee29217bb19d","Type":"ContainerStarted","Data":"08e47abb6ce4ac7cdc970b5631220273286b46f3b28f77d7080f3ce30c1d41e8"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.889207 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" event={"ID":"f25b1b4a-7676-4f7c-847a-ee29217bb19d","Type":"ContainerStarted","Data":"a232037cafc8b9ee07080910fd63f7b7b14cd7c3e073f9b59113e00824728d88"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.890912 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.893445 4966 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-j476q container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" start-of-body= Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.893486 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" podUID="f25b1b4a-7676-4f7c-847a-ee29217bb19d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.908799 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" podStartSLOduration=144.908783369 podStartE2EDuration="2m24.908783369s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:47.908527492 +0000 UTC m=+163.453597424" watchObservedRunningTime="2025-12-17 08:23:47.908783369 +0000 UTC m=+163.453853311" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.938127 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" event={"ID":"399368ce-306d-4f0b-aef7-af19955056fa","Type":"ContainerStarted","Data":"8c180e9f765db0633f727dd65a79443cec492edc255e088cdfad06c074b9e4f4"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.940404 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.941591 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.440737435 +0000 UTC m=+163.985807377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.941776 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:47 crc kubenswrapper[4966]: E1217 08:23:47.942786 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.442759981 +0000 UTC m=+163.987829923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.949384 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" event={"ID":"c3149310-e8c4-4c36-8ff6-993af0481122","Type":"ContainerStarted","Data":"fe40c94b601ed096e47bc876717270313549c734a910a76b9936fac2b7b2ece5"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.955435 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" event={"ID":"12c4901f-3d9d-40c6-a224-e31a3f0d8b97","Type":"ContainerStarted","Data":"ae84d6a13fe1df5418ccd5a90dcec16601f6453dcbd2b738b166f55547a54b00"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.983369 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6r6bw" event={"ID":"82ef7cd2-8d99-4fc4-a877-6eabfb6cd1fa","Type":"ContainerStarted","Data":"f962bd699a4d7d40b19059e1038fd36399a17d6ea0941b29c40d389169faf42a"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.985905 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" event={"ID":"c6523c2a-3325-4334-ba30-7323b1597aaf","Type":"ContainerStarted","Data":"42fe1191ae5be2fb3afc8e0b89fb0fcc39431141fc1855ea3dee78ede1517fb7"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.986654 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.987592 4966 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-f7sbq container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.987629 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" podUID="c6523c2a-3325-4334-ba30-7323b1597aaf" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.988637 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" event={"ID":"2fd94046-2a2c-45c3-9922-0962c7f7b3eb","Type":"ContainerStarted","Data":"87cf721f61d0bd3e46f4ff3d76abcf986f48046c4f84b4f89d279c4a9ed16657"} Dec 17 08:23:47 crc kubenswrapper[4966]: I1217 08:23:47.997502 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c56mr" podStartSLOduration=144.997473249 podStartE2EDuration="2m24.997473249s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:47.954609219 +0000 UTC m=+163.499679161" watchObservedRunningTime="2025-12-17 08:23:47.997473249 +0000 UTC m=+163.542543191" Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.005903 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-429md" event={"ID":"88d927f5-13ac-470d-8544-9a4aa1dc55ff","Type":"ContainerStarted","Data":"14484950df06133fc795c1b4d13715aca63fe3238b22bb6d9a59eb44a09a9981"} Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.026613 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" event={"ID":"b915a9bf-4d3d-44ae-aa8f-243283528d77","Type":"ContainerStarted","Data":"8117dd65b938a220e32de6267a6cdec60823613e3396d331bf3fce67dbecaea9"} Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.046073 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.047317 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.54730063 +0000 UTC m=+164.092370572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.067621 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cjclj" podStartSLOduration=145.067605893 podStartE2EDuration="2m25.067605893s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:47.996964955 +0000 UTC m=+163.542034897" watchObservedRunningTime="2025-12-17 08:23:48.067605893 +0000 UTC m=+163.612675835" Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.088139 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-922bk" event={"ID":"c7bbf1fa-fe65-456b-97ff-c57e2786ad73","Type":"ContainerStarted","Data":"6721e33d93d17a15a1a6329b5181ef8e2c85285fa3808ff3fa81d07dfe849d6d"} Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.111112 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktkmn" podStartSLOduration=145.111094009 podStartE2EDuration="2m25.111094009s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:48.068776606 +0000 UTC m=+163.613846548" watchObservedRunningTime="2025-12-17 08:23:48.111094009 +0000 UTC m=+163.656163951" Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.112149 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" podStartSLOduration=145.112140068 podStartE2EDuration="2m25.112140068s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:48.110306907 +0000 UTC m=+163.655376859" watchObservedRunningTime="2025-12-17 08:23:48.112140068 +0000 UTC m=+163.657210030" Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.149670 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.151770 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.651759977 +0000 UTC m=+164.196829919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.184417 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-j9jk4" podStartSLOduration=145.184402252 podStartE2EDuration="2m25.184402252s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:48.140884065 +0000 UTC m=+163.685954007" watchObservedRunningTime="2025-12-17 08:23:48.184402252 +0000 UTC m=+163.729472194" Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.224966 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" podStartSLOduration=145.224948816 podStartE2EDuration="2m25.224948816s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:48.224297898 +0000 UTC m=+163.769367830" watchObservedRunningTime="2025-12-17 08:23:48.224948816 +0000 UTC m=+163.770018758" Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.226521 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-6r6bw" podStartSLOduration=9.226514269 podStartE2EDuration="9.226514269s" podCreationTimestamp="2025-12-17 08:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:48.18503252 +0000 UTC m=+163.730102462" watchObservedRunningTime="2025-12-17 08:23:48.226514269 +0000 UTC m=+163.771584211" Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.250331 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.251561 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.751545374 +0000 UTC m=+164.296615316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.288982 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:48 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:48 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:48 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.289038 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.355753 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.356060 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.856050622 +0000 UTC m=+164.401120564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.460142 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.460601 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.96055656 +0000 UTC m=+164.505626502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.461282 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:48.961269789 +0000 UTC m=+164.506339731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.460705 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.572957 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.573512 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.073495951 +0000 UTC m=+164.618565893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.675049 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.675528 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.1755086 +0000 UTC m=+164.720578622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.776366 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.776515 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.276494951 +0000 UTC m=+164.821564893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.776674 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.777006 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.276996505 +0000 UTC m=+164.822066447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.877845 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.878201 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.37818792 +0000 UTC m=+164.923257862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:48 crc kubenswrapper[4966]: I1217 08:23:48.979426 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:48 crc kubenswrapper[4966]: E1217 08:23:48.979758 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.479746136 +0000 UTC m=+165.024816078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.084332 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.084991 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.584975866 +0000 UTC m=+165.130045808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.089808 4966 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nxfxg container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.37:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.089887 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" podUID="021d4c31-6de9-4e76-ac80-0f345b3d7f81" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.37:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.144615 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" event={"ID":"b915a9bf-4d3d-44ae-aa8f-243283528d77","Type":"ContainerStarted","Data":"ffb94a77a129a8c0533077d09040edcc64342c568c10e7565ea676ff733256bd"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.153135 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-922bk" event={"ID":"c7bbf1fa-fe65-456b-97ff-c57e2786ad73","Type":"ContainerStarted","Data":"2941397f6a3719782ab68f230b2dc07568071f230ec0d78ec3bf31a1210d8fff"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.163895 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" event={"ID":"aba2b97e-b9a0-4bc8-a480-84f64c2a0c55","Type":"ContainerStarted","Data":"fb1f5af3c54d4639002d30f2c4e3910e1dfca417c060965910b196f412be9c23"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.165961 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9f9j7" event={"ID":"8410b9e5-a83f-4bef-a99a-736c634d9e41","Type":"ContainerStarted","Data":"470408d3e02ab4297da19160a7aee352a33f7311f62c19ab58ca992c7e45fc34"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.165997 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9f9j7" event={"ID":"8410b9e5-a83f-4bef-a99a-736c634d9e41","Type":"ContainerStarted","Data":"20b4d33751851fe10cee0978c41c39a74cbe48234c5cb7e295c08f77f37eb8f6"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.166516 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.186486 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nqmws" event={"ID":"6cf617b8-bd09-49de-b188-9c86d5fafd57","Type":"ContainerStarted","Data":"621ef72651d0f32fba103bae4431b25da0f91e45812b302d861849609cb4fdfc"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.186530 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nqmws" event={"ID":"6cf617b8-bd09-49de-b188-9c86d5fafd57","Type":"ContainerStarted","Data":"8366241925fdf9bc2540b7a46df20762a7f4e4f230b933ec3844b1635965b526"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.186976 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.187250 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.687240211 +0000 UTC m=+165.232310153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.202538 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" event={"ID":"97829faf-132d-43f9-8b6a-8b6e21c3f83e","Type":"ContainerStarted","Data":"d59894cde3240e4a148e2c179868f237e8790dd06b891c3be67e921627834b97"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.207113 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" event={"ID":"4d9189ca-6d00-4596-980a-690341335c2a","Type":"ContainerStarted","Data":"579827f76261984f3357b3124ec0dbcdb11ee4dceb71e37d657857b1284b65b0"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.207156 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" event={"ID":"4d9189ca-6d00-4596-980a-690341335c2a","Type":"ContainerStarted","Data":"2afc0c7ee763e83788a5d442d4145c8e80efe4ec806fda724d957044e7b31a59"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.207266 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.216353 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" event={"ID":"46dfa510-96af-4d45-9be7-5bd9a8588a61","Type":"ContainerStarted","Data":"6479f6ecdfc99e0c7ed79e92fa54e56be521b133e557f9a4f590a6bb9216f41e"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.245926 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" event={"ID":"399368ce-306d-4f0b-aef7-af19955056fa","Type":"ContainerStarted","Data":"9cee91dd2db710afdf6f51771daef908a655be47dfeb7d4d505a09eeb7d809e6"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.253678 4966 generic.go:334] "Generic (PLEG): container finished" podID="2fd94046-2a2c-45c3-9922-0962c7f7b3eb" containerID="728253dce4b3cbc057c889c05e3635b95fd4e3899196a68421542e721c0eccec" exitCode=0 Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.253749 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" event={"ID":"2fd94046-2a2c-45c3-9922-0962c7f7b3eb","Type":"ContainerDied","Data":"728253dce4b3cbc057c889c05e3635b95fd4e3899196a68421542e721c0eccec"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.262654 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" event={"ID":"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c","Type":"ContainerStarted","Data":"03dd0839c60eb9aa19f2ad1e5c5977b922196b3dae99e97feaacb80b30c06d65"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.262882 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.264721 4966 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-km2tl container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.264760 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" podUID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.267160 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" event={"ID":"9bc07aea-e7b5-473c-957c-a5295edc2f4e","Type":"ContainerStarted","Data":"5057125f0113eedefe674fe8786bd1bfb196007e3dc353d61e0172c0bfe86551"} Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.277303 4966 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-jplds container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.277356 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" podUID="32bd318f-c091-4913-889f-9097777ceb8e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.277481 4966 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-j476q container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" start-of-body= Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.277546 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" podUID="f25b1b4a-7676-4f7c-847a-ee29217bb19d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.290368 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.290997 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:49 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:49 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:49 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.291051 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.291717 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.791700807 +0000 UTC m=+165.336770779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.316481 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-f7sbq" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.380543 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zxqjj" podStartSLOduration=146.380527651 podStartE2EDuration="2m26.380527651s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:49.26870593 +0000 UTC m=+164.813775872" watchObservedRunningTime="2025-12-17 08:23:49.380527651 +0000 UTC m=+164.925597593" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.381120 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-922bk" podStartSLOduration=146.381113777 podStartE2EDuration="2m26.381113777s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:49.38016834 +0000 UTC m=+164.925238282" watchObservedRunningTime="2025-12-17 08:23:49.381113777 +0000 UTC m=+164.926183729" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.397425 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.401737 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.901720239 +0000 UTC m=+165.446790261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.499208 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.499305 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.999290553 +0000 UTC m=+165.544360495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.499478 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.499725 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:49.999717306 +0000 UTC m=+165.544787258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.514845 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" podStartSLOduration=146.514820385 podStartE2EDuration="2m26.514820385s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:49.474904558 +0000 UTC m=+165.019974500" watchObservedRunningTime="2025-12-17 08:23:49.514820385 +0000 UTC m=+165.059890337" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.515749 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-sxhs2" podStartSLOduration=146.51574301 podStartE2EDuration="2m26.51574301s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:49.514360951 +0000 UTC m=+165.059430903" watchObservedRunningTime="2025-12-17 08:23:49.51574301 +0000 UTC m=+165.060812952" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.587347 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lr2gg" podStartSLOduration=146.587330225 podStartE2EDuration="2m26.587330225s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:49.565766927 +0000 UTC m=+165.110836869" watchObservedRunningTime="2025-12-17 08:23:49.587330225 +0000 UTC m=+165.132400167" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.608213 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.608527 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:50.108512442 +0000 UTC m=+165.653582384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.717592 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.717857 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:50.217846395 +0000 UTC m=+165.762916337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.818813 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.819282 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:50.319264837 +0000 UTC m=+165.864334779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.865510 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-9f9j7" podStartSLOduration=10.865494669 podStartE2EDuration="10.865494669s" podCreationTimestamp="2025-12-17 08:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:49.864251874 +0000 UTC m=+165.409321816" watchObservedRunningTime="2025-12-17 08:23:49.865494669 +0000 UTC m=+165.410564611" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.866808 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-8j46g" podStartSLOduration=146.866800115 podStartE2EDuration="2m26.866800115s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:49.773077976 +0000 UTC m=+165.318147918" watchObservedRunningTime="2025-12-17 08:23:49.866800115 +0000 UTC m=+165.411870057" Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.931996 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:49 crc kubenswrapper[4966]: E1217 08:23:49.932254 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:50.43224388 +0000 UTC m=+165.977313812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:49 crc kubenswrapper[4966]: I1217 08:23:49.948079 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" podStartSLOduration=146.948058658 podStartE2EDuration="2m26.948058658s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:49.93801401 +0000 UTC m=+165.483083952" watchObservedRunningTime="2025-12-17 08:23:49.948058658 +0000 UTC m=+165.493128600" Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.014467 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" podStartSLOduration=147.014452229 podStartE2EDuration="2m27.014452229s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:50.012963248 +0000 UTC m=+165.558033190" watchObservedRunningTime="2025-12-17 08:23:50.014452229 +0000 UTC m=+165.559522171" Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.032967 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.033262 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:50.533247841 +0000 UTC m=+166.078317783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.134623 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.135032 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:50.635016583 +0000 UTC m=+166.180086525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.235255 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.235539 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:50.735523989 +0000 UTC m=+166.280593931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.298740 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:50 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:50 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:50 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.299178 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.303408 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" event={"ID":"2fd94046-2a2c-45c3-9922-0962c7f7b3eb","Type":"ContainerStarted","Data":"a77e551f1177befa861717ddd00a87393d90272168dbf2292c3ad235167ee521"} Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.306448 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nqmws" event={"ID":"6cf617b8-bd09-49de-b188-9c86d5fafd57","Type":"ContainerStarted","Data":"103eb956619325079c03c593f952114dd7d84d40791dace457882a003a145a30"} Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.308170 4966 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-km2tl container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.308315 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" podUID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.339333 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.340409 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:50.840397858 +0000 UTC m=+166.385467800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.340637 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" podStartSLOduration=147.340622584 podStartE2EDuration="2m27.340622584s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:50.338453634 +0000 UTC m=+165.883523576" watchObservedRunningTime="2025-12-17 08:23:50.340622584 +0000 UTC m=+165.885692526" Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.363099 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jplds" Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.440031 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-nqmws" podStartSLOduration=147.44001273 podStartE2EDuration="2m27.44001273s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:50.391056842 +0000 UTC m=+165.936126784" watchObservedRunningTime="2025-12-17 08:23:50.44001273 +0000 UTC m=+165.985082662" Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.449244 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.450152 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:50.950133991 +0000 UTC m=+166.495203943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.550526 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.550857 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.050822682 +0000 UTC m=+166.595892624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.651724 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.651964 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.151939837 +0000 UTC m=+166.697009779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.652052 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.652422 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.15241008 +0000 UTC m=+166.697480022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.753197 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.753549 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.253535084 +0000 UTC m=+166.798605026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.861432 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.861786 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.361774466 +0000 UTC m=+166.906844408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:50 crc kubenswrapper[4966]: I1217 08:23:50.962198 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:50 crc kubenswrapper[4966]: E1217 08:23:50.963241 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.463224928 +0000 UTC m=+167.008294870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.063096 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.063474 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.563461478 +0000 UTC m=+167.108531420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.163974 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.164139 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.66411398 +0000 UTC m=+167.209183922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.164316 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.164662 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.664651514 +0000 UTC m=+167.209721456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.265899 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.266086 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.766049565 +0000 UTC m=+167.311119507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.266185 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.266457 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.766446867 +0000 UTC m=+167.311516799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.285377 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:51 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:51 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:51 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.285658 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.307918 4966 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-j476q container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.307969 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" podUID="f25b1b4a-7676-4f7c-847a-ee29217bb19d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.324123 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" event={"ID":"46dfa510-96af-4d45-9be7-5bd9a8588a61","Type":"ContainerStarted","Data":"e5fb681b43080f11aab094f16dab1af247e234729f774c2f1e5e098811e31256"} Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.324160 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" event={"ID":"46dfa510-96af-4d45-9be7-5bd9a8588a61","Type":"ContainerStarted","Data":"0de5b91cdd3166345173da8b2fb94931a0917a438c4ff8da916c31da8483ef78"} Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.359711 4966 patch_prober.go:28] interesting pod/downloads-7954f5f757-r5fmr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.359722 4966 patch_prober.go:28] interesting pod/downloads-7954f5f757-r5fmr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.359760 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r5fmr" podUID="091f07a2-acd7-4d1e-b242-220c47716c25" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.359773 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r5fmr" podUID="091f07a2-acd7-4d1e-b242-220c47716c25" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.367192 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.368371 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.868355912 +0000 UTC m=+167.413425854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.469108 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.469533 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:51.969511738 +0000 UTC m=+167.514581760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.540496 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.540719 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.570680 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.570862 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.070838098 +0000 UTC m=+167.615908040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.571384 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.571813 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.071798105 +0000 UTC m=+167.616868037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.673194 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.673565 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.173540825 +0000 UTC m=+167.718610767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.774409 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.774761 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.274746162 +0000 UTC m=+167.819816104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.815165 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.815221 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.816372 4966 patch_prober.go:28] interesting pod/console-f9d7485db-b998v container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.816409 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-b998v" podUID="992684e8-e50b-44e7-b30c-7169f4877695" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.876079 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.876288 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.376254956 +0000 UTC m=+167.921324898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.876585 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.877459 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.37745062 +0000 UTC m=+167.922520562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.884095 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zvg6n"] Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.884978 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.935120 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.984250 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.984544 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-catalog-content\") pod \"community-operators-zvg6n\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.984607 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psjr8\" (UniqueName: \"kubernetes.io/projected/d5c75559-5110-464f-893d-421f55a5318f-kube-api-access-psjr8\") pod \"community-operators-zvg6n\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:51 crc kubenswrapper[4966]: I1217 08:23:51.984635 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-utilities\") pod \"community-operators-zvg6n\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:51 crc kubenswrapper[4966]: E1217 08:23:51.984735 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.484718695 +0000 UTC m=+168.029788637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.025639 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zvg6n"] Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.085338 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psjr8\" (UniqueName: \"kubernetes.io/projected/d5c75559-5110-464f-893d-421f55a5318f-kube-api-access-psjr8\") pod \"community-operators-zvg6n\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.085373 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g57dd"] Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.085400 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-utilities\") pod \"community-operators-zvg6n\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.085464 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-catalog-content\") pod \"community-operators-zvg6n\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.085490 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.085801 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.585790047 +0000 UTC m=+168.130859989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.086216 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.086515 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-utilities\") pod \"community-operators-zvg6n\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.086744 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-catalog-content\") pod \"community-operators-zvg6n\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.114219 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.164705 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g57dd"] Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.196128 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.196357 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.696338403 +0000 UTC m=+168.241408345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.196508 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-utilities\") pod \"certified-operators-g57dd\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.196528 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psjr8\" (UniqueName: \"kubernetes.io/projected/d5c75559-5110-464f-893d-421f55a5318f-kube-api-access-psjr8\") pod \"community-operators-zvg6n\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.196703 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.196748 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-catalog-content\") pod \"certified-operators-g57dd\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.196788 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjxmn\" (UniqueName: \"kubernetes.io/projected/763b5661-27a9-4a61-9de9-17b7f8832720-kube-api-access-zjxmn\") pod \"certified-operators-g57dd\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.197080 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.697069503 +0000 UTC m=+168.242139445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.200152 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.288340 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6spc2"] Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.290394 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.291058 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.299691 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.300055 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.800026968 +0000 UTC m=+168.345096910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.300665 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.301088 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.801077268 +0000 UTC m=+168.346147210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.301927 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-catalog-content\") pod \"certified-operators-g57dd\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.302046 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjxmn\" (UniqueName: \"kubernetes.io/projected/763b5661-27a9-4a61-9de9-17b7f8832720-kube-api-access-zjxmn\") pod \"certified-operators-g57dd\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.302217 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-utilities\") pod \"certified-operators-g57dd\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.303047 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-utilities\") pod \"certified-operators-g57dd\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.303560 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-catalog-content\") pod \"certified-operators-g57dd\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.306611 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:52 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:52 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:52 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.306904 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.332447 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.377817 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" event={"ID":"46dfa510-96af-4d45-9be7-5bd9a8588a61","Type":"ContainerStarted","Data":"180976e50968d2c026bf24c79a69976ab043a90c4d6c2bae71614cf3c962eb6a"} Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.410625 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjxmn\" (UniqueName: \"kubernetes.io/projected/763b5661-27a9-4a61-9de9-17b7f8832720-kube-api-access-zjxmn\") pod \"certified-operators-g57dd\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.411368 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.411525 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6spc2"] Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.411763 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-utilities\") pod \"community-operators-6spc2\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.411885 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrtms\" (UniqueName: \"kubernetes.io/projected/690f46fd-f538-4eec-af73-3c16a9ad7a82-kube-api-access-vrtms\") pod \"community-operators-6spc2\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.412049 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-catalog-content\") pod \"community-operators-6spc2\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.413143 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:52.913127215 +0000 UTC m=+168.458197157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.413320 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.515652 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-82qfl"] Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.516804 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.519620 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrtms\" (UniqueName: \"kubernetes.io/projected/690f46fd-f538-4eec-af73-3c16a9ad7a82-kube-api-access-vrtms\") pod \"community-operators-6spc2\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.519660 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-catalog-content\") pod \"community-operators-6spc2\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.519718 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.519753 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-utilities\") pod \"community-operators-6spc2\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.520102 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-utilities\") pod \"community-operators-6spc2\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.535000 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-catalog-content\") pod \"community-operators-6spc2\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.535361 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:53.035348964 +0000 UTC m=+168.580418906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.562925 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-82qfl"] Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.625533 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.625739 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-catalog-content\") pod \"certified-operators-82qfl\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.625825 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km9js\" (UniqueName: \"kubernetes.io/projected/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-kube-api-access-km9js\") pod \"certified-operators-82qfl\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.625852 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-utilities\") pod \"certified-operators-82qfl\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.625955 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:53.125941716 +0000 UTC m=+168.671011658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.633299 4966 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.648499 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrtms\" (UniqueName: \"kubernetes.io/projected/690f46fd-f538-4eec-af73-3c16a9ad7a82-kube-api-access-vrtms\") pod \"community-operators-6spc2\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.660018 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" podStartSLOduration=13.66000364 podStartE2EDuration="13.66000364s" podCreationTimestamp="2025-12-17 08:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:52.651437493 +0000 UTC m=+168.196507435" watchObservedRunningTime="2025-12-17 08:23:52.66000364 +0000 UTC m=+168.205073572" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.663138 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.726655 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-utilities\") pod \"certified-operators-82qfl\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.726891 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km9js\" (UniqueName: \"kubernetes.io/projected/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-kube-api-access-km9js\") pod \"certified-operators-82qfl\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.726924 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.726951 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-catalog-content\") pod \"certified-operators-82qfl\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.727296 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-catalog-content\") pod \"certified-operators-82qfl\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.727518 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-utilities\") pod \"certified-operators-82qfl\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.733385 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:53.233366434 +0000 UTC m=+168.778436376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.795243 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km9js\" (UniqueName: \"kubernetes.io/projected/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-kube-api-access-km9js\") pod \"certified-operators-82qfl\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.828162 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.828662 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:53.328646266 +0000 UTC m=+168.873716208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.894681 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:23:52 crc kubenswrapper[4966]: I1217 08:23:52.931684 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:52 crc kubenswrapper[4966]: E1217 08:23:52.932055 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-17 08:23:53.432043364 +0000 UTC m=+168.977113306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sfbnp" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.023826 4966 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-17T08:23:52.633329481Z","Handler":null,"Name":""} Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.033428 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:53 crc kubenswrapper[4966]: E1217 08:23:53.033820 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-17 08:23:53.533793575 +0000 UTC m=+169.078863517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.051497 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zvg6n"] Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.070646 4966 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.070707 4966 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.134814 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.222825 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.226356 4966 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.226390 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.241148 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.241186 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.252726 4966 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-km2tl container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.253000 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" podUID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.252737 4966 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-km2tl container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.253343 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" podUID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.299607 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:53 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:53 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:53 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.299666 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.309220 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.405734 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvg6n" event={"ID":"d5c75559-5110-464f-893d-421f55a5318f","Type":"ContainerStarted","Data":"13cdb043859552b8d74247cd5f8391eb672839fd8c05fe514e8517a5726cbb69"} Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.423312 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.459244 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.460023 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.464278 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.464500 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.533372 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.652557 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.652623 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.750240 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sfbnp\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.759045 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.759121 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.759272 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.860133 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.860618 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:53 crc kubenswrapper[4966]: I1217 08:23:53.973613 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.012401 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gllj5"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.013597 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.025048 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.040962 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gllj5"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.075494 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g57dd"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.097099 4966 patch_prober.go:28] interesting pod/apiserver-76f77b778f-922bk container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]log ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]etcd ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/generic-apiserver-start-informers ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/max-in-flight-filter ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 17 08:23:54 crc kubenswrapper[4966]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 17 08:23:54 crc kubenswrapper[4966]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/project.openshift.io-projectcache ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/openshift.io-startinformers ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 17 08:23:54 crc kubenswrapper[4966]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 17 08:23:54 crc kubenswrapper[4966]: livez check failed Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.097156 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-922bk" podUID="c7bbf1fa-fe65-456b-97ff-c57e2786ad73" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.097695 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.175583 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8jh9\" (UniqueName: \"kubernetes.io/projected/c6569dae-8fd2-4536-a645-51eed6946e41-kube-api-access-g8jh9\") pod \"redhat-marketplace-gllj5\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.175664 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-catalog-content\") pod \"redhat-marketplace-gllj5\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.175704 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-utilities\") pod \"redhat-marketplace-gllj5\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.186382 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.243622 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6spc2"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.255216 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-82qfl"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.276662 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-catalog-content\") pod \"redhat-marketplace-gllj5\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.276728 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-utilities\") pod \"redhat-marketplace-gllj5\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.276773 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8jh9\" (UniqueName: \"kubernetes.io/projected/c6569dae-8fd2-4536-a645-51eed6946e41-kube-api-access-g8jh9\") pod \"redhat-marketplace-gllj5\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.277801 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-catalog-content\") pod \"redhat-marketplace-gllj5\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.278021 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-utilities\") pod \"redhat-marketplace-gllj5\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.297056 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:54 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:54 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:54 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.297302 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.316862 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8jh9\" (UniqueName: \"kubernetes.io/projected/c6569dae-8fd2-4536-a645-51eed6946e41-kube-api-access-g8jh9\") pod \"redhat-marketplace-gllj5\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.363644 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.402895 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-75zxk"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.403919 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.427042 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-75zxk"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.452266 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6spc2" event={"ID":"690f46fd-f538-4eec-af73-3c16a9ad7a82","Type":"ContainerStarted","Data":"14655b4fc903fa55bf50918726947761e88f324c8f2ffa2f4ef435e0aed4997e"} Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.475851 4966 generic.go:334] "Generic (PLEG): container finished" podID="d5c75559-5110-464f-893d-421f55a5318f" containerID="5efbceefa6eeb8f08adae2d3981dcf5e6ffdf658883bf68aa64219e10f229b2d" exitCode=0 Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.475965 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvg6n" event={"ID":"d5c75559-5110-464f-893d-421f55a5318f","Type":"ContainerDied","Data":"5efbceefa6eeb8f08adae2d3981dcf5e6ffdf658883bf68aa64219e10f229b2d"} Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.485911 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sfbnp"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.486926 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.490332 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4pk9\" (UniqueName: \"kubernetes.io/projected/9ddfa746-5822-4c08-a190-4e2f7edd41ff-kube-api-access-r4pk9\") pod \"redhat-marketplace-75zxk\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.490370 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-utilities\") pod \"redhat-marketplace-75zxk\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.490409 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-catalog-content\") pod \"redhat-marketplace-75zxk\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.491244 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g57dd" event={"ID":"763b5661-27a9-4a61-9de9-17b7f8832720","Type":"ContainerStarted","Data":"8c394aa8bee6f039909efc58bc38da4d9a5889634e46a4a3707eb3923db04ec7"} Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.534175 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82qfl" event={"ID":"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7","Type":"ContainerStarted","Data":"1b22aded81a264ab60065fbfb6060885241fe1b9250e1c96af002d5183277133"} Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.593086 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-utilities\") pod \"redhat-marketplace-75zxk\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.593367 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-catalog-content\") pod \"redhat-marketplace-75zxk\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.593524 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4pk9\" (UniqueName: \"kubernetes.io/projected/9ddfa746-5822-4c08-a190-4e2f7edd41ff-kube-api-access-r4pk9\") pod \"redhat-marketplace-75zxk\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.595184 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-utilities\") pod \"redhat-marketplace-75zxk\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.595605 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-catalog-content\") pod \"redhat-marketplace-75zxk\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.609950 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4pk9\" (UniqueName: \"kubernetes.io/projected/9ddfa746-5822-4c08-a190-4e2f7edd41ff-kube-api-access-r4pk9\") pod \"redhat-marketplace-75zxk\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.694194 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.735375 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.842788 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.856408 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gllj5"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.990714 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ssxxq"] Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.991709 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:54 crc kubenswrapper[4966]: I1217 08:23:54.996057 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.000322 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ssxxq"] Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.100862 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-utilities\") pod \"redhat-operators-ssxxq\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.100983 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8dpr\" (UniqueName: \"kubernetes.io/projected/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-kube-api-access-n8dpr\") pod \"redhat-operators-ssxxq\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.101181 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-catalog-content\") pod \"redhat-operators-ssxxq\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.202245 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-catalog-content\") pod \"redhat-operators-ssxxq\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.202321 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-utilities\") pod \"redhat-operators-ssxxq\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.202395 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8dpr\" (UniqueName: \"kubernetes.io/projected/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-kube-api-access-n8dpr\") pod \"redhat-operators-ssxxq\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.203141 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-catalog-content\") pod \"redhat-operators-ssxxq\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.203151 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-utilities\") pod \"redhat-operators-ssxxq\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.228531 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8dpr\" (UniqueName: \"kubernetes.io/projected/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-kube-api-access-n8dpr\") pod \"redhat-operators-ssxxq\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.284635 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:55 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:55 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:55 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.284702 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.305429 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.401446 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xlrlx"] Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.402583 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.425729 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlrlx"] Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.505443 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-utilities\") pod \"redhat-operators-xlrlx\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.505519 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-catalog-content\") pod \"redhat-operators-xlrlx\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.505551 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gprb\" (UniqueName: \"kubernetes.io/projected/88efab12-8605-40ee-9912-992901670b00-kube-api-access-8gprb\") pod \"redhat-operators-xlrlx\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.606640 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-utilities\") pod \"redhat-operators-xlrlx\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.606710 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-catalog-content\") pod \"redhat-operators-xlrlx\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.606752 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gprb\" (UniqueName: \"kubernetes.io/projected/88efab12-8605-40ee-9912-992901670b00-kube-api-access-8gprb\") pod \"redhat-operators-xlrlx\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.607319 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-utilities\") pod \"redhat-operators-xlrlx\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.610152 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-catalog-content\") pod \"redhat-operators-xlrlx\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.614558 4966 generic.go:334] "Generic (PLEG): container finished" podID="9bc07aea-e7b5-473c-957c-a5295edc2f4e" containerID="5057125f0113eedefe674fe8786bd1bfb196007e3dc353d61e0172c0bfe86551" exitCode=0 Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.614650 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" event={"ID":"9bc07aea-e7b5-473c-957c-a5295edc2f4e","Type":"ContainerDied","Data":"5057125f0113eedefe674fe8786bd1bfb196007e3dc353d61e0172c0bfe86551"} Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.641975 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gprb\" (UniqueName: \"kubernetes.io/projected/88efab12-8605-40ee-9912-992901670b00-kube-api-access-8gprb\") pod \"redhat-operators-xlrlx\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.642828 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae","Type":"ContainerStarted","Data":"e8ffc2b0e3affd3f31ca87abd4bb799148963a0146141aeb25f45398ab5bcca2"} Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.689954 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6spc2" event={"ID":"690f46fd-f538-4eec-af73-3c16a9ad7a82","Type":"ContainerStarted","Data":"eb5c6f9b8272468856c4f85965a3f22ad08d5477035392f641e3a463733b647f"} Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.703147 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g57dd" event={"ID":"763b5661-27a9-4a61-9de9-17b7f8832720","Type":"ContainerStarted","Data":"21001dacd7269f5387581569ad4f2b486efd54f2c8bf3ea69bea0b53c1b716b8"} Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.706843 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" event={"ID":"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a","Type":"ContainerStarted","Data":"92c1092c1be301494b7b99760bd966c0d4cb19dd3c92dbffc42b4e552a8fcc4c"} Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.719418 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:23:55 crc kubenswrapper[4966]: I1217 08:23:55.748232 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82qfl" event={"ID":"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7","Type":"ContainerStarted","Data":"8536a8e48521411ae5e2bbac058342f6670992d1b6f112601ffdb7d6b653029d"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.057041 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-75zxk"] Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.071695 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ssxxq"] Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.293192 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:56 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:56 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:56 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.293563 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.293748 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.294415 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.304355 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.304639 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.311798 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.335177 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/699f9b84-6aa2-4feb-adbb-2b6932411a94-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"699f9b84-6aa2-4feb-adbb-2b6932411a94\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.335300 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/699f9b84-6aa2-4feb-adbb-2b6932411a94-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"699f9b84-6aa2-4feb-adbb-2b6932411a94\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:23:56 crc kubenswrapper[4966]: E1217 08:23:56.366071 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc82ff2ce_39a3_48bb_a4d3_69cfe999b8a2.slice/crio-d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c.scope\": RecentStats: unable to find data in memory cache]" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.436343 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/699f9b84-6aa2-4feb-adbb-2b6932411a94-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"699f9b84-6aa2-4feb-adbb-2b6932411a94\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.436549 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/699f9b84-6aa2-4feb-adbb-2b6932411a94-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"699f9b84-6aa2-4feb-adbb-2b6932411a94\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.436944 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/699f9b84-6aa2-4feb-adbb-2b6932411a94-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"699f9b84-6aa2-4feb-adbb-2b6932411a94\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.437330 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlrlx"] Dec 17 08:23:56 crc kubenswrapper[4966]: W1217 08:23:56.458054 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88efab12_8605_40ee_9912_992901670b00.slice/crio-04bd4ce9a132c460ce070bab4c10ae0e32c808371e6ca3328a05aaab53c2a696 WatchSource:0}: Error finding container 04bd4ce9a132c460ce070bab4c10ae0e32c808371e6ca3328a05aaab53c2a696: Status 404 returned error can't find the container with id 04bd4ce9a132c460ce070bab4c10ae0e32c808371e6ca3328a05aaab53c2a696 Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.480710 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/699f9b84-6aa2-4feb-adbb-2b6932411a94-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"699f9b84-6aa2-4feb-adbb-2b6932411a94\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.544679 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.552532 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-922bk" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.689345 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.816267 4966 generic.go:334] "Generic (PLEG): container finished" podID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerID="d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c" exitCode=0 Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.816353 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ssxxq" event={"ID":"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2","Type":"ContainerDied","Data":"d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.816376 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ssxxq" event={"ID":"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2","Type":"ContainerStarted","Data":"ed993ff61ed1ac93f202bec5556728b8bfcb8a49285e1828217a394fd8957d56"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.829683 4966 generic.go:334] "Generic (PLEG): container finished" podID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerID="8536a8e48521411ae5e2bbac058342f6670992d1b6f112601ffdb7d6b653029d" exitCode=0 Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.888258 4966 generic.go:334] "Generic (PLEG): container finished" podID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerID="22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5" exitCode=0 Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.889116 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82qfl" event={"ID":"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7","Type":"ContainerDied","Data":"8536a8e48521411ae5e2bbac058342f6670992d1b6f112601ffdb7d6b653029d"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.889149 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75zxk" event={"ID":"9ddfa746-5822-4c08-a190-4e2f7edd41ff","Type":"ContainerDied","Data":"22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.889162 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75zxk" event={"ID":"9ddfa746-5822-4c08-a190-4e2f7edd41ff","Type":"ContainerStarted","Data":"1bd2ea1c89f0bd7cab80382d2eb4ecaba681ff82d16bf8f96f94a0db69d2fb80"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.925166 4966 generic.go:334] "Generic (PLEG): container finished" podID="ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae" containerID="2db165bc966beeb222ba7d6fada762f4b1b5db80ceefe380b4349581078c0396" exitCode=0 Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.925263 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae","Type":"ContainerDied","Data":"2db165bc966beeb222ba7d6fada762f4b1b5db80ceefe380b4349581078c0396"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.928721 4966 generic.go:334] "Generic (PLEG): container finished" podID="c6569dae-8fd2-4536-a645-51eed6946e41" containerID="c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943" exitCode=0 Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.928796 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gllj5" event={"ID":"c6569dae-8fd2-4536-a645-51eed6946e41","Type":"ContainerDied","Data":"c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.928820 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gllj5" event={"ID":"c6569dae-8fd2-4536-a645-51eed6946e41","Type":"ContainerStarted","Data":"2dbcaad8e29b9b345684e76bacc17e8b225a3b469d5d0afddcea4b177c3befb6"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.955556 4966 generic.go:334] "Generic (PLEG): container finished" podID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerID="eb5c6f9b8272468856c4f85965a3f22ad08d5477035392f641e3a463733b647f" exitCode=0 Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.955628 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6spc2" event={"ID":"690f46fd-f538-4eec-af73-3c16a9ad7a82","Type":"ContainerDied","Data":"eb5c6f9b8272468856c4f85965a3f22ad08d5477035392f641e3a463733b647f"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.976138 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlrlx" event={"ID":"88efab12-8605-40ee-9912-992901670b00","Type":"ContainerStarted","Data":"e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.976195 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlrlx" event={"ID":"88efab12-8605-40ee-9912-992901670b00","Type":"ContainerStarted","Data":"04bd4ce9a132c460ce070bab4c10ae0e32c808371e6ca3328a05aaab53c2a696"} Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.992700 4966 generic.go:334] "Generic (PLEG): container finished" podID="763b5661-27a9-4a61-9de9-17b7f8832720" containerID="21001dacd7269f5387581569ad4f2b486efd54f2c8bf3ea69bea0b53c1b716b8" exitCode=0 Dec 17 08:23:56 crc kubenswrapper[4966]: I1217 08:23:56.992909 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g57dd" event={"ID":"763b5661-27a9-4a61-9de9-17b7f8832720","Type":"ContainerDied","Data":"21001dacd7269f5387581569ad4f2b486efd54f2c8bf3ea69bea0b53c1b716b8"} Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.001276 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" event={"ID":"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a","Type":"ContainerStarted","Data":"d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2"} Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.001316 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.036915 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" podStartSLOduration=154.036893582 podStartE2EDuration="2m34.036893582s" podCreationTimestamp="2025-12-17 08:21:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:57.035922334 +0000 UTC m=+172.580992276" watchObservedRunningTime="2025-12-17 08:23:57.036893582 +0000 UTC m=+172.581963524" Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.392094 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:57 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:57 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:57 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.392153 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.569014 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.816397 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.995624 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bc07aea-e7b5-473c-957c-a5295edc2f4e-secret-volume\") pod \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.995728 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bc07aea-e7b5-473c-957c-a5295edc2f4e-config-volume\") pod \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.995818 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjknh\" (UniqueName: \"kubernetes.io/projected/9bc07aea-e7b5-473c-957c-a5295edc2f4e-kube-api-access-bjknh\") pod \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\" (UID: \"9bc07aea-e7b5-473c-957c-a5295edc2f4e\") " Dec 17 08:23:57 crc kubenswrapper[4966]: I1217 08:23:57.997102 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc07aea-e7b5-473c-957c-a5295edc2f4e-config-volume" (OuterVolumeSpecName: "config-volume") pod "9bc07aea-e7b5-473c-957c-a5295edc2f4e" (UID: "9bc07aea-e7b5-473c-957c-a5295edc2f4e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.014145 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bc07aea-e7b5-473c-957c-a5295edc2f4e-kube-api-access-bjknh" (OuterVolumeSpecName: "kube-api-access-bjknh") pod "9bc07aea-e7b5-473c-957c-a5295edc2f4e" (UID: "9bc07aea-e7b5-473c-957c-a5295edc2f4e"). InnerVolumeSpecName "kube-api-access-bjknh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.014732 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bc07aea-e7b5-473c-957c-a5295edc2f4e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9bc07aea-e7b5-473c-957c-a5295edc2f4e" (UID: "9bc07aea-e7b5-473c-957c-a5295edc2f4e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.063855 4966 generic.go:334] "Generic (PLEG): container finished" podID="88efab12-8605-40ee-9912-992901670b00" containerID="e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6" exitCode=0 Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.064151 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlrlx" event={"ID":"88efab12-8605-40ee-9912-992901670b00","Type":"ContainerDied","Data":"e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6"} Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.066112 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.066107 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4" event={"ID":"9bc07aea-e7b5-473c-957c-a5295edc2f4e","Type":"ContainerDied","Data":"c061623f0d05889623a8db49b4e7330bd34d798f9ff934112212fe7035f5f58b"} Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.066163 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c061623f0d05889623a8db49b4e7330bd34d798f9ff934112212fe7035f5f58b" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.087914 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"699f9b84-6aa2-4feb-adbb-2b6932411a94","Type":"ContainerStarted","Data":"cbf68c64cb4510bee62234e37513a6cd0fc143db2559740778536dd1e2624e0a"} Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.096973 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bc07aea-e7b5-473c-957c-a5295edc2f4e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.097001 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bc07aea-e7b5-473c-957c-a5295edc2f4e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.097009 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjknh\" (UniqueName: \"kubernetes.io/projected/9bc07aea-e7b5-473c-957c-a5295edc2f4e-kube-api-access-bjknh\") on node \"crc\" DevicePath \"\"" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.292700 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-9f9j7" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.300225 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:58 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:58 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:58 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.300286 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.822646 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.942928 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kubelet-dir\") pod \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\" (UID: \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\") " Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.942986 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kube-api-access\") pod \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\" (UID: \"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae\") " Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.943085 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae" (UID: "ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.943211 4966 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:23:58 crc kubenswrapper[4966]: I1217 08:23:58.963763 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae" (UID: "ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:23:59 crc kubenswrapper[4966]: I1217 08:23:59.044335 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 17 08:23:59 crc kubenswrapper[4966]: I1217 08:23:59.144258 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"699f9b84-6aa2-4feb-adbb-2b6932411a94","Type":"ContainerStarted","Data":"dda6c1a73ca6e0ecb1282fd0adeaa3e13f53b14a6db3c1934b92ad88294293c4"} Dec 17 08:23:59 crc kubenswrapper[4966]: I1217 08:23:59.153894 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae","Type":"ContainerDied","Data":"e8ffc2b0e3affd3f31ca87abd4bb799148963a0146141aeb25f45398ab5bcca2"} Dec 17 08:23:59 crc kubenswrapper[4966]: I1217 08:23:59.153935 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8ffc2b0e3affd3f31ca87abd4bb799148963a0146141aeb25f45398ab5bcca2" Dec 17 08:23:59 crc kubenswrapper[4966]: I1217 08:23:59.153999 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 17 08:23:59 crc kubenswrapper[4966]: I1217 08:23:59.168088 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.168071269 podStartE2EDuration="3.168071269s" podCreationTimestamp="2025-12-17 08:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:23:59.155782168 +0000 UTC m=+174.700852110" watchObservedRunningTime="2025-12-17 08:23:59.168071269 +0000 UTC m=+174.713141211" Dec 17 08:23:59 crc kubenswrapper[4966]: I1217 08:23:59.283575 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:23:59 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:23:59 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:23:59 crc kubenswrapper[4966]: healthz check failed Dec 17 08:23:59 crc kubenswrapper[4966]: I1217 08:23:59.283624 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:24:00 crc kubenswrapper[4966]: I1217 08:24:00.201437 4966 generic.go:334] "Generic (PLEG): container finished" podID="699f9b84-6aa2-4feb-adbb-2b6932411a94" containerID="dda6c1a73ca6e0ecb1282fd0adeaa3e13f53b14a6db3c1934b92ad88294293c4" exitCode=0 Dec 17 08:24:00 crc kubenswrapper[4966]: I1217 08:24:00.201475 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"699f9b84-6aa2-4feb-adbb-2b6932411a94","Type":"ContainerDied","Data":"dda6c1a73ca6e0ecb1282fd0adeaa3e13f53b14a6db3c1934b92ad88294293c4"} Dec 17 08:24:00 crc kubenswrapper[4966]: I1217 08:24:00.282573 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:24:00 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:24:00 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:24:00 crc kubenswrapper[4966]: healthz check failed Dec 17 08:24:00 crc kubenswrapper[4966]: I1217 08:24:00.282637 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.282889 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:24:01 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:24:01 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:24:01 crc kubenswrapper[4966]: healthz check failed Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.282942 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.439148 4966 patch_prober.go:28] interesting pod/downloads-7954f5f757-r5fmr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.439205 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r5fmr" podUID="091f07a2-acd7-4d1e-b242-220c47716c25" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.439483 4966 patch_prober.go:28] interesting pod/downloads-7954f5f757-r5fmr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.439987 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r5fmr" podUID="091f07a2-acd7-4d1e-b242-220c47716c25" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.840709 4966 patch_prober.go:28] interesting pod/console-f9d7485db-b998v container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.840762 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-b998v" podUID="992684e8-e50b-44e7-b30c-7169f4877695" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.867629 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.951725 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/699f9b84-6aa2-4feb-adbb-2b6932411a94-kube-api-access\") pod \"699f9b84-6aa2-4feb-adbb-2b6932411a94\" (UID: \"699f9b84-6aa2-4feb-adbb-2b6932411a94\") " Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.951781 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/699f9b84-6aa2-4feb-adbb-2b6932411a94-kubelet-dir\") pod \"699f9b84-6aa2-4feb-adbb-2b6932411a94\" (UID: \"699f9b84-6aa2-4feb-adbb-2b6932411a94\") " Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.952171 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/699f9b84-6aa2-4feb-adbb-2b6932411a94-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "699f9b84-6aa2-4feb-adbb-2b6932411a94" (UID: "699f9b84-6aa2-4feb-adbb-2b6932411a94"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:24:01 crc kubenswrapper[4966]: I1217 08:24:01.978013 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/699f9b84-6aa2-4feb-adbb-2b6932411a94-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "699f9b84-6aa2-4feb-adbb-2b6932411a94" (UID: "699f9b84-6aa2-4feb-adbb-2b6932411a94"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:24:02 crc kubenswrapper[4966]: I1217 08:24:02.053510 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/699f9b84-6aa2-4feb-adbb-2b6932411a94-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 17 08:24:02 crc kubenswrapper[4966]: I1217 08:24:02.053536 4966 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/699f9b84-6aa2-4feb-adbb-2b6932411a94-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:24:02 crc kubenswrapper[4966]: I1217 08:24:02.235766 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"699f9b84-6aa2-4feb-adbb-2b6932411a94","Type":"ContainerDied","Data":"cbf68c64cb4510bee62234e37513a6cd0fc143db2559740778536dd1e2624e0a"} Dec 17 08:24:02 crc kubenswrapper[4966]: I1217 08:24:02.236061 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 17 08:24:02 crc kubenswrapper[4966]: I1217 08:24:02.236173 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbf68c64cb4510bee62234e37513a6cd0fc143db2559740778536dd1e2624e0a" Dec 17 08:24:02 crc kubenswrapper[4966]: I1217 08:24:02.292502 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:24:02 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:24:02 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:24:02 crc kubenswrapper[4966]: healthz check failed Dec 17 08:24:02 crc kubenswrapper[4966]: I1217 08:24:02.292554 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:24:03 crc kubenswrapper[4966]: I1217 08:24:03.261306 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:24:03 crc kubenswrapper[4966]: I1217 08:24:03.293943 4966 patch_prober.go:28] interesting pod/router-default-5444994796-587gm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 17 08:24:03 crc kubenswrapper[4966]: [-]has-synced failed: reason withheld Dec 17 08:24:03 crc kubenswrapper[4966]: [+]process-running ok Dec 17 08:24:03 crc kubenswrapper[4966]: healthz check failed Dec 17 08:24:03 crc kubenswrapper[4966]: I1217 08:24:03.293991 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-587gm" podUID="8b1849b5-04d6-4550-bd09-0f69c712b283" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:24:04 crc kubenswrapper[4966]: I1217 08:24:04.284096 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:24:04 crc kubenswrapper[4966]: I1217 08:24:04.287210 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-587gm" Dec 17 08:24:11 crc kubenswrapper[4966]: I1217 08:24:11.360959 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-r5fmr" Dec 17 08:24:11 crc kubenswrapper[4966]: I1217 08:24:11.831123 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:24:11 crc kubenswrapper[4966]: I1217 08:24:11.838214 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:24:14 crc kubenswrapper[4966]: I1217 08:24:14.048116 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:24:16 crc kubenswrapper[4966]: I1217 08:24:16.807684 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:24:16 crc kubenswrapper[4966]: I1217 08:24:16.808017 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:24:23 crc kubenswrapper[4966]: I1217 08:24:23.221107 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gw7bx" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.090617 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 17 08:24:34 crc kubenswrapper[4966]: E1217 08:24:34.091404 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc07aea-e7b5-473c-957c-a5295edc2f4e" containerName="collect-profiles" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.091420 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc07aea-e7b5-473c-957c-a5295edc2f4e" containerName="collect-profiles" Dec 17 08:24:34 crc kubenswrapper[4966]: E1217 08:24:34.091439 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="699f9b84-6aa2-4feb-adbb-2b6932411a94" containerName="pruner" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.091447 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="699f9b84-6aa2-4feb-adbb-2b6932411a94" containerName="pruner" Dec 17 08:24:34 crc kubenswrapper[4966]: E1217 08:24:34.091463 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae" containerName="pruner" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.091471 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae" containerName="pruner" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.091587 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bc07aea-e7b5-473c-957c-a5295edc2f4e" containerName="collect-profiles" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.091603 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef6aa374-e417-4b0d-8eb3-23bcfda9e2ae" containerName="pruner" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.091611 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="699f9b84-6aa2-4feb-adbb-2b6932411a94" containerName="pruner" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.092105 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.094532 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.094684 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.095398 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.130514 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f747c74d-ea98-497e-84f8-0efc92408d51-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f747c74d-ea98-497e-84f8-0efc92408d51\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.130554 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f747c74d-ea98-497e-84f8-0efc92408d51-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f747c74d-ea98-497e-84f8-0efc92408d51\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.231520 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f747c74d-ea98-497e-84f8-0efc92408d51-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f747c74d-ea98-497e-84f8-0efc92408d51\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.231577 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f747c74d-ea98-497e-84f8-0efc92408d51-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f747c74d-ea98-497e-84f8-0efc92408d51\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.231687 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f747c74d-ea98-497e-84f8-0efc92408d51-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f747c74d-ea98-497e-84f8-0efc92408d51\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.249549 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f747c74d-ea98-497e-84f8-0efc92408d51-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f747c74d-ea98-497e-84f8-0efc92408d51\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:34 crc kubenswrapper[4966]: I1217 08:24:34.423817 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:37 crc kubenswrapper[4966]: E1217 08:24:37.876028 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 17 08:24:37 crc kubenswrapper[4966]: E1217 08:24:37.877833 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-km9js,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-82qfl_openshift-marketplace(95ac14e7-1ba1-4f8e-96c8-46c89f1732a7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 17 08:24:37 crc kubenswrapper[4966]: E1217 08:24:37.880011 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-82qfl" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" Dec 17 08:24:37 crc kubenswrapper[4966]: E1217 08:24:37.907958 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 17 08:24:37 crc kubenswrapper[4966]: E1217 08:24:37.908102 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zjxmn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-g57dd_openshift-marketplace(763b5661-27a9-4a61-9de9-17b7f8832720): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 17 08:24:37 crc kubenswrapper[4966]: E1217 08:24:37.909425 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-g57dd" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" Dec 17 08:24:39 crc kubenswrapper[4966]: E1217 08:24:39.418061 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-82qfl" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" Dec 17 08:24:39 crc kubenswrapper[4966]: E1217 08:24:39.418084 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-g57dd" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" Dec 17 08:24:39 crc kubenswrapper[4966]: E1217 08:24:39.497575 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 17 08:24:39 crc kubenswrapper[4966]: E1217 08:24:39.497834 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-psjr8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zvg6n_openshift-marketplace(d5c75559-5110-464f-893d-421f55a5318f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 17 08:24:39 crc kubenswrapper[4966]: E1217 08:24:39.498992 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zvg6n" podUID="d5c75559-5110-464f-893d-421f55a5318f" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.072285 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.072934 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.086259 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.143131 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-var-lock\") pod \"installer-9-crc\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.143283 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.143327 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/771b39e4-c3a9-4363-b57f-a429cc16157e-kube-api-access\") pod \"installer-9-crc\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.246558 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-var-lock\") pod \"installer-9-crc\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.246956 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.246655 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-var-lock\") pod \"installer-9-crc\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.247104 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/771b39e4-c3a9-4363-b57f-a429cc16157e-kube-api-access\") pod \"installer-9-crc\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.247160 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.264533 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/771b39e4-c3a9-4363-b57f-a429cc16157e-kube-api-access\") pod \"installer-9-crc\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:40 crc kubenswrapper[4966]: I1217 08:24:40.409737 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:24:43 crc kubenswrapper[4966]: E1217 08:24:43.574641 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zvg6n" podUID="d5c75559-5110-464f-893d-421f55a5318f" Dec 17 08:24:43 crc kubenswrapper[4966]: E1217 08:24:43.642406 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 17 08:24:43 crc kubenswrapper[4966]: E1217 08:24:43.642545 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n8dpr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-ssxxq_openshift-marketplace(c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 17 08:24:43 crc kubenswrapper[4966]: E1217 08:24:43.643865 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-ssxxq" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" Dec 17 08:24:45 crc kubenswrapper[4966]: E1217 08:24:45.900122 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-ssxxq" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.077966 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.078452 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g8jh9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-gllj5_openshift-marketplace(c6569dae-8fd2-4536-a645-51eed6946e41): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.079784 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-gllj5" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.308713 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 17 08:24:46 crc kubenswrapper[4966]: W1217 08:24:46.313182 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podf747c74d_ea98_497e_84f8_0efc92408d51.slice/crio-c378f87c30f4c5ef0cec2bb132112ad31ffdd98293e8d6e6633adc6043b25dd1 WatchSource:0}: Error finding container c378f87c30f4c5ef0cec2bb132112ad31ffdd98293e8d6e6633adc6043b25dd1: Status 404 returned error can't find the container with id c378f87c30f4c5ef0cec2bb132112ad31ffdd98293e8d6e6633adc6043b25dd1 Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.360746 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.662684 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.663152 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8gprb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xlrlx_openshift-marketplace(88efab12-8605-40ee-9912-992901670b00): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.664335 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xlrlx" podUID="88efab12-8605-40ee-9912-992901670b00" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.706372 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.706600 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4pk9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-75zxk_openshift-marketplace(9ddfa746-5822-4c08-a190-4e2f7edd41ff): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.708047 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-75zxk" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.792663 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.792890 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vrtms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6spc2_openshift-marketplace(690f46fd-f538-4eec-af73-3c16a9ad7a82): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.794131 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6spc2" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.809001 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.809068 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.809118 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.809846 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.809964 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532" gracePeriod=600 Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.974345 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f747c74d-ea98-497e-84f8-0efc92408d51","Type":"ContainerStarted","Data":"084f906faed8028e3db27feb3358ac6f8ce85f04177ddb5fdf261b1381eba153"} Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.974415 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f747c74d-ea98-497e-84f8-0efc92408d51","Type":"ContainerStarted","Data":"c378f87c30f4c5ef0cec2bb132112ad31ffdd98293e8d6e6633adc6043b25dd1"} Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.976598 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532" exitCode=0 Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.976710 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532"} Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.979891 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"771b39e4-c3a9-4363-b57f-a429cc16157e","Type":"ContainerStarted","Data":"297409173e166c1e3d780663674d32edbb3e402d8f37354a8d54bb02bdcc01bb"} Dec 17 08:24:46 crc kubenswrapper[4966]: I1217 08:24:46.979923 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"771b39e4-c3a9-4363-b57f-a429cc16157e","Type":"ContainerStarted","Data":"31ce221e8475f59b458301dd2cd6d659f57d1d265ccac64766e52be50e3bd317"} Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.982054 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xlrlx" podUID="88efab12-8605-40ee-9912-992901670b00" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.982177 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-gllj5" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.982594 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-75zxk" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" Dec 17 08:24:46 crc kubenswrapper[4966]: E1217 08:24:46.982706 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6spc2" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" Dec 17 08:24:47 crc kubenswrapper[4966]: I1217 08:24:47.024726 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=13.024699544 podStartE2EDuration="13.024699544s" podCreationTimestamp="2025-12-17 08:24:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:24:46.987004476 +0000 UTC m=+222.532074418" watchObservedRunningTime="2025-12-17 08:24:47.024699544 +0000 UTC m=+222.569769496" Dec 17 08:24:47 crc kubenswrapper[4966]: I1217 08:24:47.093676 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=7.093661576 podStartE2EDuration="7.093661576s" podCreationTimestamp="2025-12-17 08:24:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:24:47.059249397 +0000 UTC m=+222.604319339" watchObservedRunningTime="2025-12-17 08:24:47.093661576 +0000 UTC m=+222.638731598" Dec 17 08:24:47 crc kubenswrapper[4966]: I1217 08:24:47.989246 4966 generic.go:334] "Generic (PLEG): container finished" podID="f747c74d-ea98-497e-84f8-0efc92408d51" containerID="084f906faed8028e3db27feb3358ac6f8ce85f04177ddb5fdf261b1381eba153" exitCode=0 Dec 17 08:24:47 crc kubenswrapper[4966]: I1217 08:24:47.989339 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f747c74d-ea98-497e-84f8-0efc92408d51","Type":"ContainerDied","Data":"084f906faed8028e3db27feb3358ac6f8ce85f04177ddb5fdf261b1381eba153"} Dec 17 08:24:47 crc kubenswrapper[4966]: I1217 08:24:47.991581 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"af9614c017f65ecd41597b9880f18912c6650c9ea6e2d776d690cac78b26db56"} Dec 17 08:24:49 crc kubenswrapper[4966]: I1217 08:24:49.280236 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:49 crc kubenswrapper[4966]: I1217 08:24:49.451918 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f747c74d-ea98-497e-84f8-0efc92408d51-kubelet-dir\") pod \"f747c74d-ea98-497e-84f8-0efc92408d51\" (UID: \"f747c74d-ea98-497e-84f8-0efc92408d51\") " Dec 17 08:24:49 crc kubenswrapper[4966]: I1217 08:24:49.452186 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f747c74d-ea98-497e-84f8-0efc92408d51-kube-api-access\") pod \"f747c74d-ea98-497e-84f8-0efc92408d51\" (UID: \"f747c74d-ea98-497e-84f8-0efc92408d51\") " Dec 17 08:24:49 crc kubenswrapper[4966]: I1217 08:24:49.451990 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f747c74d-ea98-497e-84f8-0efc92408d51-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f747c74d-ea98-497e-84f8-0efc92408d51" (UID: "f747c74d-ea98-497e-84f8-0efc92408d51"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:24:49 crc kubenswrapper[4966]: I1217 08:24:49.452551 4966 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f747c74d-ea98-497e-84f8-0efc92408d51-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:24:49 crc kubenswrapper[4966]: I1217 08:24:49.457632 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f747c74d-ea98-497e-84f8-0efc92408d51-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f747c74d-ea98-497e-84f8-0efc92408d51" (UID: "f747c74d-ea98-497e-84f8-0efc92408d51"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:24:49 crc kubenswrapper[4966]: I1217 08:24:49.554209 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f747c74d-ea98-497e-84f8-0efc92408d51-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 17 08:24:50 crc kubenswrapper[4966]: I1217 08:24:50.001825 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f747c74d-ea98-497e-84f8-0efc92408d51","Type":"ContainerDied","Data":"c378f87c30f4c5ef0cec2bb132112ad31ffdd98293e8d6e6633adc6043b25dd1"} Dec 17 08:24:50 crc kubenswrapper[4966]: I1217 08:24:50.001884 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c378f87c30f4c5ef0cec2bb132112ad31ffdd98293e8d6e6633adc6043b25dd1" Dec 17 08:24:50 crc kubenswrapper[4966]: I1217 08:24:50.001905 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 17 08:24:58 crc kubenswrapper[4966]: I1217 08:24:58.038592 4966 generic.go:334] "Generic (PLEG): container finished" podID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerID="7def3cf5a493813273e4a5483fd4e1fbee44aef0d46fcedea889a45827b38a28" exitCode=0 Dec 17 08:24:58 crc kubenswrapper[4966]: I1217 08:24:58.038814 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82qfl" event={"ID":"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7","Type":"ContainerDied","Data":"7def3cf5a493813273e4a5483fd4e1fbee44aef0d46fcedea889a45827b38a28"} Dec 17 08:24:59 crc kubenswrapper[4966]: I1217 08:24:59.045235 4966 generic.go:334] "Generic (PLEG): container finished" podID="d5c75559-5110-464f-893d-421f55a5318f" containerID="52c0d1eb699e068249ad478405cc98ae555299fd794c4128ad677584b5acfa58" exitCode=0 Dec 17 08:24:59 crc kubenswrapper[4966]: I1217 08:24:59.045332 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvg6n" event={"ID":"d5c75559-5110-464f-893d-421f55a5318f","Type":"ContainerDied","Data":"52c0d1eb699e068249ad478405cc98ae555299fd794c4128ad677584b5acfa58"} Dec 17 08:24:59 crc kubenswrapper[4966]: I1217 08:24:59.053082 4966 generic.go:334] "Generic (PLEG): container finished" podID="763b5661-27a9-4a61-9de9-17b7f8832720" containerID="233bf7829ec27fc201d8c721d7ac53acf2b3f927e348fd828b6f99d16d99c949" exitCode=0 Dec 17 08:24:59 crc kubenswrapper[4966]: I1217 08:24:59.053168 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g57dd" event={"ID":"763b5661-27a9-4a61-9de9-17b7f8832720","Type":"ContainerDied","Data":"233bf7829ec27fc201d8c721d7ac53acf2b3f927e348fd828b6f99d16d99c949"} Dec 17 08:24:59 crc kubenswrapper[4966]: I1217 08:24:59.056417 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82qfl" event={"ID":"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7","Type":"ContainerStarted","Data":"d68762197838c30c68e7faf2ce29e6904eed1d7706059ac2bd8950a37472e3bb"} Dec 17 08:24:59 crc kubenswrapper[4966]: I1217 08:24:59.098941 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-82qfl" podStartSLOduration=4.18045124 podStartE2EDuration="1m7.098924028s" podCreationTimestamp="2025-12-17 08:23:52 +0000 UTC" firstStartedPulling="2025-12-17 08:23:55.765204408 +0000 UTC m=+171.310274350" lastFinishedPulling="2025-12-17 08:24:58.683677196 +0000 UTC m=+234.228747138" observedRunningTime="2025-12-17 08:24:59.080225868 +0000 UTC m=+234.625295850" watchObservedRunningTime="2025-12-17 08:24:59.098924028 +0000 UTC m=+234.643993990" Dec 17 08:25:00 crc kubenswrapper[4966]: I1217 08:25:00.064290 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvg6n" event={"ID":"d5c75559-5110-464f-893d-421f55a5318f","Type":"ContainerStarted","Data":"c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8"} Dec 17 08:25:00 crc kubenswrapper[4966]: I1217 08:25:00.066996 4966 generic.go:334] "Generic (PLEG): container finished" podID="c6569dae-8fd2-4536-a645-51eed6946e41" containerID="f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432" exitCode=0 Dec 17 08:25:00 crc kubenswrapper[4966]: I1217 08:25:00.067123 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gllj5" event={"ID":"c6569dae-8fd2-4536-a645-51eed6946e41","Type":"ContainerDied","Data":"f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432"} Dec 17 08:25:00 crc kubenswrapper[4966]: I1217 08:25:00.071449 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g57dd" event={"ID":"763b5661-27a9-4a61-9de9-17b7f8832720","Type":"ContainerStarted","Data":"12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3"} Dec 17 08:25:00 crc kubenswrapper[4966]: I1217 08:25:00.085599 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zvg6n" podStartSLOduration=3.864565277 podStartE2EDuration="1m9.085585825s" podCreationTimestamp="2025-12-17 08:23:51 +0000 UTC" firstStartedPulling="2025-12-17 08:23:54.486696255 +0000 UTC m=+170.031766187" lastFinishedPulling="2025-12-17 08:24:59.707716783 +0000 UTC m=+235.252786735" observedRunningTime="2025-12-17 08:25:00.083098887 +0000 UTC m=+235.628168849" watchObservedRunningTime="2025-12-17 08:25:00.085585825 +0000 UTC m=+235.630655757" Dec 17 08:25:00 crc kubenswrapper[4966]: I1217 08:25:00.127185 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g57dd" podStartSLOduration=5.334451005 podStartE2EDuration="1m9.12716725s" podCreationTimestamp="2025-12-17 08:23:51 +0000 UTC" firstStartedPulling="2025-12-17 08:23:55.704558555 +0000 UTC m=+171.249628497" lastFinishedPulling="2025-12-17 08:24:59.4972748 +0000 UTC m=+235.042344742" observedRunningTime="2025-12-17 08:25:00.126316977 +0000 UTC m=+235.671386939" watchObservedRunningTime="2025-12-17 08:25:00.12716725 +0000 UTC m=+235.672237182" Dec 17 08:25:01 crc kubenswrapper[4966]: I1217 08:25:01.083282 4966 generic.go:334] "Generic (PLEG): container finished" podID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerID="25f99d75dd3e2117a7d430cf9e1d409a7d6cec3b3803cd53b0322f7192651f0c" exitCode=0 Dec 17 08:25:01 crc kubenswrapper[4966]: I1217 08:25:01.083380 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6spc2" event={"ID":"690f46fd-f538-4eec-af73-3c16a9ad7a82","Type":"ContainerDied","Data":"25f99d75dd3e2117a7d430cf9e1d409a7d6cec3b3803cd53b0322f7192651f0c"} Dec 17 08:25:01 crc kubenswrapper[4966]: I1217 08:25:01.091540 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gllj5" event={"ID":"c6569dae-8fd2-4536-a645-51eed6946e41","Type":"ContainerStarted","Data":"7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a"} Dec 17 08:25:01 crc kubenswrapper[4966]: I1217 08:25:01.130212 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gllj5" podStartSLOduration=4.327787121 podStartE2EDuration="1m8.130196383s" podCreationTimestamp="2025-12-17 08:23:53 +0000 UTC" firstStartedPulling="2025-12-17 08:23:56.947455011 +0000 UTC m=+172.492524953" lastFinishedPulling="2025-12-17 08:25:00.749864253 +0000 UTC m=+236.294934215" observedRunningTime="2025-12-17 08:25:01.127935811 +0000 UTC m=+236.673005753" watchObservedRunningTime="2025-12-17 08:25:01.130196383 +0000 UTC m=+236.675266325" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.098896 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6spc2" event={"ID":"690f46fd-f538-4eec-af73-3c16a9ad7a82","Type":"ContainerStarted","Data":"58b618abfbf2185db3acdd64fb4fe10d31d3c4b5137bb39d009d3855ee45d7a5"} Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.100927 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlrlx" event={"ID":"88efab12-8605-40ee-9912-992901670b00","Type":"ContainerStarted","Data":"ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427"} Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.119304 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6spc2" podStartSLOduration=4.254859239 podStartE2EDuration="1m10.119288817s" podCreationTimestamp="2025-12-17 08:23:52 +0000 UTC" firstStartedPulling="2025-12-17 08:23:55.691607407 +0000 UTC m=+171.236677349" lastFinishedPulling="2025-12-17 08:25:01.556036985 +0000 UTC m=+237.101106927" observedRunningTime="2025-12-17 08:25:02.118262929 +0000 UTC m=+237.663332881" watchObservedRunningTime="2025-12-17 08:25:02.119288817 +0000 UTC m=+237.664358749" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.201127 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.201776 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.275393 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.417099 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.417169 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.469951 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.664293 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.664351 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.895793 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.896083 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:25:02 crc kubenswrapper[4966]: I1217 08:25:02.959045 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:25:03 crc kubenswrapper[4966]: I1217 08:25:03.106891 4966 generic.go:334] "Generic (PLEG): container finished" podID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerID="7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362" exitCode=0 Dec 17 08:25:03 crc kubenswrapper[4966]: I1217 08:25:03.106949 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75zxk" event={"ID":"9ddfa746-5822-4c08-a190-4e2f7edd41ff","Type":"ContainerDied","Data":"7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362"} Dec 17 08:25:03 crc kubenswrapper[4966]: I1217 08:25:03.111135 4966 generic.go:334] "Generic (PLEG): container finished" podID="88efab12-8605-40ee-9912-992901670b00" containerID="ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427" exitCode=0 Dec 17 08:25:03 crc kubenswrapper[4966]: I1217 08:25:03.111328 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlrlx" event={"ID":"88efab12-8605-40ee-9912-992901670b00","Type":"ContainerDied","Data":"ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427"} Dec 17 08:25:03 crc kubenswrapper[4966]: I1217 08:25:03.166594 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:25:03 crc kubenswrapper[4966]: I1217 08:25:03.698313 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6spc2" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerName="registry-server" probeResult="failure" output=< Dec 17 08:25:03 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:25:03 crc kubenswrapper[4966]: > Dec 17 08:25:04 crc kubenswrapper[4966]: I1217 08:25:04.150087 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:25:04 crc kubenswrapper[4966]: I1217 08:25:04.364301 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:25:04 crc kubenswrapper[4966]: I1217 08:25:04.364417 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:25:04 crc kubenswrapper[4966]: I1217 08:25:04.406744 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:25:05 crc kubenswrapper[4966]: I1217 08:25:05.161328 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:25:05 crc kubenswrapper[4966]: I1217 08:25:05.904806 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-82qfl"] Dec 17 08:25:06 crc kubenswrapper[4966]: I1217 08:25:06.129714 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlrlx" event={"ID":"88efab12-8605-40ee-9912-992901670b00","Type":"ContainerStarted","Data":"237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf"} Dec 17 08:25:06 crc kubenswrapper[4966]: I1217 08:25:06.133524 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ssxxq" event={"ID":"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2","Type":"ContainerStarted","Data":"b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9"} Dec 17 08:25:06 crc kubenswrapper[4966]: I1217 08:25:06.135465 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75zxk" event={"ID":"9ddfa746-5822-4c08-a190-4e2f7edd41ff","Type":"ContainerStarted","Data":"f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948"} Dec 17 08:25:06 crc kubenswrapper[4966]: I1217 08:25:06.135492 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-82qfl" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerName="registry-server" containerID="cri-o://d68762197838c30c68e7faf2ce29e6904eed1d7706059ac2bd8950a37472e3bb" gracePeriod=2 Dec 17 08:25:06 crc kubenswrapper[4966]: I1217 08:25:06.160017 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xlrlx" podStartSLOduration=2.550516585 podStartE2EDuration="1m11.15999812s" podCreationTimestamp="2025-12-17 08:23:55 +0000 UTC" firstStartedPulling="2025-12-17 08:23:56.984486318 +0000 UTC m=+172.529556260" lastFinishedPulling="2025-12-17 08:25:05.593967853 +0000 UTC m=+241.139037795" observedRunningTime="2025-12-17 08:25:06.1563078 +0000 UTC m=+241.701377772" watchObservedRunningTime="2025-12-17 08:25:06.15999812 +0000 UTC m=+241.705068072" Dec 17 08:25:06 crc kubenswrapper[4966]: I1217 08:25:06.177748 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-75zxk" podStartSLOduration=3.531633521 podStartE2EDuration="1m12.177731234s" podCreationTimestamp="2025-12-17 08:23:54 +0000 UTC" firstStartedPulling="2025-12-17 08:23:56.897194567 +0000 UTC m=+172.442264509" lastFinishedPulling="2025-12-17 08:25:05.54329229 +0000 UTC m=+241.088362222" observedRunningTime="2025-12-17 08:25:06.176358186 +0000 UTC m=+241.721428138" watchObservedRunningTime="2025-12-17 08:25:06.177731234 +0000 UTC m=+241.722801186" Dec 17 08:25:07 crc kubenswrapper[4966]: I1217 08:25:07.155070 4966 generic.go:334] "Generic (PLEG): container finished" podID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerID="d68762197838c30c68e7faf2ce29e6904eed1d7706059ac2bd8950a37472e3bb" exitCode=0 Dec 17 08:25:07 crc kubenswrapper[4966]: I1217 08:25:07.155109 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82qfl" event={"ID":"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7","Type":"ContainerDied","Data":"d68762197838c30c68e7faf2ce29e6904eed1d7706059ac2bd8950a37472e3bb"} Dec 17 08:25:07 crc kubenswrapper[4966]: I1217 08:25:07.947672 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.117073 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-catalog-content\") pod \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.117162 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km9js\" (UniqueName: \"kubernetes.io/projected/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-kube-api-access-km9js\") pod \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.117205 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-utilities\") pod \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\" (UID: \"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7\") " Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.118102 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-utilities" (OuterVolumeSpecName: "utilities") pod "95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" (UID: "95ac14e7-1ba1-4f8e-96c8-46c89f1732a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.125481 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-kube-api-access-km9js" (OuterVolumeSpecName: "kube-api-access-km9js") pod "95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" (UID: "95ac14e7-1ba1-4f8e-96c8-46c89f1732a7"). InnerVolumeSpecName "kube-api-access-km9js". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.164221 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82qfl" event={"ID":"95ac14e7-1ba1-4f8e-96c8-46c89f1732a7","Type":"ContainerDied","Data":"1b22aded81a264ab60065fbfb6060885241fe1b9250e1c96af002d5183277133"} Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.164515 4966 scope.go:117] "RemoveContainer" containerID="d68762197838c30c68e7faf2ce29e6904eed1d7706059ac2bd8950a37472e3bb" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.164285 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82qfl" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.179325 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" (UID: "95ac14e7-1ba1-4f8e-96c8-46c89f1732a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.180153 4966 scope.go:117] "RemoveContainer" containerID="7def3cf5a493813273e4a5483fd4e1fbee44aef0d46fcedea889a45827b38a28" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.194175 4966 scope.go:117] "RemoveContainer" containerID="8536a8e48521411ae5e2bbac058342f6670992d1b6f112601ffdb7d6b653029d" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.218950 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.218984 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km9js\" (UniqueName: \"kubernetes.io/projected/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-kube-api-access-km9js\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.218998 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.490538 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-82qfl"] Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.494884 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-82qfl"] Dec 17 08:25:08 crc kubenswrapper[4966]: I1217 08:25:08.837006 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" path="/var/lib/kubelet/pods/95ac14e7-1ba1-4f8e-96c8-46c89f1732a7/volumes" Dec 17 08:25:09 crc kubenswrapper[4966]: I1217 08:25:09.172114 4966 generic.go:334] "Generic (PLEG): container finished" podID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerID="b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9" exitCode=0 Dec 17 08:25:09 crc kubenswrapper[4966]: I1217 08:25:09.172173 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ssxxq" event={"ID":"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2","Type":"ContainerDied","Data":"b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9"} Dec 17 08:25:12 crc kubenswrapper[4966]: I1217 08:25:12.484567 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:25:12 crc kubenswrapper[4966]: I1217 08:25:12.762021 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:25:12 crc kubenswrapper[4966]: I1217 08:25:12.812705 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:25:13 crc kubenswrapper[4966]: I1217 08:25:13.301300 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6spc2"] Dec 17 08:25:14 crc kubenswrapper[4966]: I1217 08:25:14.225778 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6spc2" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerName="registry-server" containerID="cri-o://58b618abfbf2185db3acdd64fb4fe10d31d3c4b5137bb39d009d3855ee45d7a5" gracePeriod=2 Dec 17 08:25:14 crc kubenswrapper[4966]: I1217 08:25:14.226112 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ssxxq" event={"ID":"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2","Type":"ContainerStarted","Data":"ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168"} Dec 17 08:25:14 crc kubenswrapper[4966]: I1217 08:25:14.736120 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:25:14 crc kubenswrapper[4966]: I1217 08:25:14.736445 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:25:14 crc kubenswrapper[4966]: I1217 08:25:14.787904 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:25:14 crc kubenswrapper[4966]: I1217 08:25:14.807579 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ssxxq" podStartSLOduration=4.412157099 podStartE2EDuration="1m20.807558999s" podCreationTimestamp="2025-12-17 08:23:54 +0000 UTC" firstStartedPulling="2025-12-17 08:23:56.821964211 +0000 UTC m=+172.367034153" lastFinishedPulling="2025-12-17 08:25:13.217366111 +0000 UTC m=+248.762436053" observedRunningTime="2025-12-17 08:25:14.252009797 +0000 UTC m=+249.797079749" watchObservedRunningTime="2025-12-17 08:25:14.807558999 +0000 UTC m=+250.352628941" Dec 17 08:25:15 crc kubenswrapper[4966]: I1217 08:25:15.231815 4966 generic.go:334] "Generic (PLEG): container finished" podID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerID="58b618abfbf2185db3acdd64fb4fe10d31d3c4b5137bb39d009d3855ee45d7a5" exitCode=0 Dec 17 08:25:15 crc kubenswrapper[4966]: I1217 08:25:15.232512 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6spc2" event={"ID":"690f46fd-f538-4eec-af73-3c16a9ad7a82","Type":"ContainerDied","Data":"58b618abfbf2185db3acdd64fb4fe10d31d3c4b5137bb39d009d3855ee45d7a5"} Dec 17 08:25:15 crc kubenswrapper[4966]: I1217 08:25:15.274385 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:25:15 crc kubenswrapper[4966]: I1217 08:25:15.306588 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:25:15 crc kubenswrapper[4966]: I1217 08:25:15.307074 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:25:15 crc kubenswrapper[4966]: I1217 08:25:15.719740 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:25:15 crc kubenswrapper[4966]: I1217 08:25:15.719796 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:25:15 crc kubenswrapper[4966]: I1217 08:25:15.756931 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.119716 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.238600 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6spc2" event={"ID":"690f46fd-f538-4eec-af73-3c16a9ad7a82","Type":"ContainerDied","Data":"14655b4fc903fa55bf50918726947761e88f324c8f2ffa2f4ef435e0aed4997e"} Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.238655 4966 scope.go:117] "RemoveContainer" containerID="58b618abfbf2185db3acdd64fb4fe10d31d3c4b5137bb39d009d3855ee45d7a5" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.238659 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6spc2" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.253092 4966 scope.go:117] "RemoveContainer" containerID="25f99d75dd3e2117a7d430cf9e1d409a7d6cec3b3803cd53b0322f7192651f0c" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.267369 4966 scope.go:117] "RemoveContainer" containerID="eb5c6f9b8272468856c4f85965a3f22ad08d5477035392f641e3a463733b647f" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.278071 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.278251 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-catalog-content\") pod \"690f46fd-f538-4eec-af73-3c16a9ad7a82\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.278356 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrtms\" (UniqueName: \"kubernetes.io/projected/690f46fd-f538-4eec-af73-3c16a9ad7a82-kube-api-access-vrtms\") pod \"690f46fd-f538-4eec-af73-3c16a9ad7a82\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.278848 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-utilities\") pod \"690f46fd-f538-4eec-af73-3c16a9ad7a82\" (UID: \"690f46fd-f538-4eec-af73-3c16a9ad7a82\") " Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.280810 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-utilities" (OuterVolumeSpecName: "utilities") pod "690f46fd-f538-4eec-af73-3c16a9ad7a82" (UID: "690f46fd-f538-4eec-af73-3c16a9ad7a82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.289035 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690f46fd-f538-4eec-af73-3c16a9ad7a82-kube-api-access-vrtms" (OuterVolumeSpecName: "kube-api-access-vrtms") pod "690f46fd-f538-4eec-af73-3c16a9ad7a82" (UID: "690f46fd-f538-4eec-af73-3c16a9ad7a82"). InnerVolumeSpecName "kube-api-access-vrtms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.337327 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "690f46fd-f538-4eec-af73-3c16a9ad7a82" (UID: "690f46fd-f538-4eec-af73-3c16a9ad7a82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.380109 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.380143 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrtms\" (UniqueName: \"kubernetes.io/projected/690f46fd-f538-4eec-af73-3c16a9ad7a82-kube-api-access-vrtms\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.380154 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/690f46fd-f538-4eec-af73-3c16a9ad7a82-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.410606 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ssxxq" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerName="registry-server" probeResult="failure" output=< Dec 17 08:25:16 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:25:16 crc kubenswrapper[4966]: > Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.566517 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6spc2"] Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.570451 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6spc2"] Dec 17 08:25:16 crc kubenswrapper[4966]: I1217 08:25:16.835686 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" path="/var/lib/kubelet/pods/690f46fd-f538-4eec-af73-3c16a9ad7a82/volumes" Dec 17 08:25:18 crc kubenswrapper[4966]: I1217 08:25:18.876000 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nxfxg"] Dec 17 08:25:19 crc kubenswrapper[4966]: I1217 08:25:19.700595 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-75zxk"] Dec 17 08:25:19 crc kubenswrapper[4966]: I1217 08:25:19.701036 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-75zxk" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerName="registry-server" containerID="cri-o://f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948" gracePeriod=2 Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.017688 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.140688 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4pk9\" (UniqueName: \"kubernetes.io/projected/9ddfa746-5822-4c08-a190-4e2f7edd41ff-kube-api-access-r4pk9\") pod \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.140998 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-utilities\") pod \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.141123 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-catalog-content\") pod \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\" (UID: \"9ddfa746-5822-4c08-a190-4e2f7edd41ff\") " Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.142899 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-utilities" (OuterVolumeSpecName: "utilities") pod "9ddfa746-5822-4c08-a190-4e2f7edd41ff" (UID: "9ddfa746-5822-4c08-a190-4e2f7edd41ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.145331 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ddfa746-5822-4c08-a190-4e2f7edd41ff-kube-api-access-r4pk9" (OuterVolumeSpecName: "kube-api-access-r4pk9") pod "9ddfa746-5822-4c08-a190-4e2f7edd41ff" (UID: "9ddfa746-5822-4c08-a190-4e2f7edd41ff"). InnerVolumeSpecName "kube-api-access-r4pk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.165992 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ddfa746-5822-4c08-a190-4e2f7edd41ff" (UID: "9ddfa746-5822-4c08-a190-4e2f7edd41ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.242329 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4pk9\" (UniqueName: \"kubernetes.io/projected/9ddfa746-5822-4c08-a190-4e2f7edd41ff-kube-api-access-r4pk9\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.242367 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.242381 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ddfa746-5822-4c08-a190-4e2f7edd41ff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.261667 4966 generic.go:334] "Generic (PLEG): container finished" podID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerID="f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948" exitCode=0 Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.261714 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75zxk" event={"ID":"9ddfa746-5822-4c08-a190-4e2f7edd41ff","Type":"ContainerDied","Data":"f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948"} Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.261736 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-75zxk" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.261752 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75zxk" event={"ID":"9ddfa746-5822-4c08-a190-4e2f7edd41ff","Type":"ContainerDied","Data":"1bd2ea1c89f0bd7cab80382d2eb4ecaba681ff82d16bf8f96f94a0db69d2fb80"} Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.261776 4966 scope.go:117] "RemoveContainer" containerID="f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.284860 4966 scope.go:117] "RemoveContainer" containerID="7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.299529 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-75zxk"] Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.307722 4966 scope.go:117] "RemoveContainer" containerID="22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.311014 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-75zxk"] Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.312711 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlrlx"] Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.313145 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xlrlx" podUID="88efab12-8605-40ee-9912-992901670b00" containerName="registry-server" containerID="cri-o://237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf" gracePeriod=2 Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.334766 4966 scope.go:117] "RemoveContainer" containerID="f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948" Dec 17 08:25:20 crc kubenswrapper[4966]: E1217 08:25:20.335334 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948\": container with ID starting with f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948 not found: ID does not exist" containerID="f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.335381 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948"} err="failed to get container status \"f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948\": rpc error: code = NotFound desc = could not find container \"f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948\": container with ID starting with f74d01dc2ae280adcf0f992021edac94b095fbf2fb173eb9388a860e34e5a948 not found: ID does not exist" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.335427 4966 scope.go:117] "RemoveContainer" containerID="7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362" Dec 17 08:25:20 crc kubenswrapper[4966]: E1217 08:25:20.335713 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362\": container with ID starting with 7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362 not found: ID does not exist" containerID="7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.335753 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362"} err="failed to get container status \"7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362\": rpc error: code = NotFound desc = could not find container \"7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362\": container with ID starting with 7197f66596636355a98f3364ba609f4543dd99940254ed953f608b592ed3c362 not found: ID does not exist" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.335782 4966 scope.go:117] "RemoveContainer" containerID="22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5" Dec 17 08:25:20 crc kubenswrapper[4966]: E1217 08:25:20.336378 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5\": container with ID starting with 22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5 not found: ID does not exist" containerID="22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.336405 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5"} err="failed to get container status \"22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5\": rpc error: code = NotFound desc = could not find container \"22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5\": container with ID starting with 22ca62a1b8485de2738a161fc76024bcabb4a3a7c2e9937ebb4c691013f5dba5 not found: ID does not exist" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.621652 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.748249 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-catalog-content\") pod \"88efab12-8605-40ee-9912-992901670b00\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.748308 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gprb\" (UniqueName: \"kubernetes.io/projected/88efab12-8605-40ee-9912-992901670b00-kube-api-access-8gprb\") pod \"88efab12-8605-40ee-9912-992901670b00\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.748380 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-utilities\") pod \"88efab12-8605-40ee-9912-992901670b00\" (UID: \"88efab12-8605-40ee-9912-992901670b00\") " Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.749479 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-utilities" (OuterVolumeSpecName: "utilities") pod "88efab12-8605-40ee-9912-992901670b00" (UID: "88efab12-8605-40ee-9912-992901670b00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.754790 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88efab12-8605-40ee-9912-992901670b00-kube-api-access-8gprb" (OuterVolumeSpecName: "kube-api-access-8gprb") pod "88efab12-8605-40ee-9912-992901670b00" (UID: "88efab12-8605-40ee-9912-992901670b00"). InnerVolumeSpecName "kube-api-access-8gprb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.838140 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" path="/var/lib/kubelet/pods/9ddfa746-5822-4c08-a190-4e2f7edd41ff/volumes" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.849745 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gprb\" (UniqueName: \"kubernetes.io/projected/88efab12-8605-40ee-9912-992901670b00-kube-api-access-8gprb\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.849774 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.862812 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88efab12-8605-40ee-9912-992901670b00" (UID: "88efab12-8605-40ee-9912-992901670b00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:20 crc kubenswrapper[4966]: I1217 08:25:20.950522 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88efab12-8605-40ee-9912-992901670b00-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.145463 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g57dd"] Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.145787 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g57dd" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" containerName="registry-server" containerID="cri-o://12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3" gracePeriod=30 Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.157787 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zvg6n"] Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.158129 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zvg6n" podUID="d5c75559-5110-464f-893d-421f55a5318f" containerName="registry-server" containerID="cri-o://c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8" gracePeriod=30 Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.171407 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-km2tl"] Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.171735 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" podUID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" containerName="marketplace-operator" containerID="cri-o://03dd0839c60eb9aa19f2ad1e5c5977b922196b3dae99e97feaacb80b30c06d65" gracePeriod=30 Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.188348 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gllj5"] Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.188643 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gllj5" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" containerName="registry-server" containerID="cri-o://7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a" gracePeriod=30 Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.201531 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ssxxq"] Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.201802 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ssxxq" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerName="registry-server" containerID="cri-o://ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168" gracePeriod=30 Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207391 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-grzxs"] Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207634 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207656 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207668 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerName="extract-content" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207678 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerName="extract-content" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207690 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207698 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207710 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerName="extract-content" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207718 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerName="extract-content" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207727 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207734 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207743 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88efab12-8605-40ee-9912-992901670b00" containerName="extract-content" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207750 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="88efab12-8605-40ee-9912-992901670b00" containerName="extract-content" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207763 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f747c74d-ea98-497e-84f8-0efc92408d51" containerName="pruner" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207770 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f747c74d-ea98-497e-84f8-0efc92408d51" containerName="pruner" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207782 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88efab12-8605-40ee-9912-992901670b00" containerName="extract-utilities" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207792 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="88efab12-8605-40ee-9912-992901670b00" containerName="extract-utilities" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207805 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerName="extract-utilities" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207813 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerName="extract-utilities" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207820 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerName="extract-utilities" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207827 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerName="extract-utilities" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207838 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerName="extract-utilities" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207845 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerName="extract-utilities" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.207853 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerName="extract-content" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.207860 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerName="extract-content" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.208014 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88efab12-8605-40ee-9912-992901670b00" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.208025 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="88efab12-8605-40ee-9912-992901670b00" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.208152 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="690f46fd-f538-4eec-af73-3c16a9ad7a82" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.208169 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f747c74d-ea98-497e-84f8-0efc92408d51" containerName="pruner" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.208181 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ddfa746-5822-4c08-a190-4e2f7edd41ff" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.208193 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ac14e7-1ba1-4f8e-96c8-46c89f1732a7" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.208200 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="88efab12-8605-40ee-9912-992901670b00" containerName="registry-server" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.208866 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.219894 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-grzxs"] Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.269587 4966 generic.go:334] "Generic (PLEG): container finished" podID="88efab12-8605-40ee-9912-992901670b00" containerID="237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf" exitCode=0 Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.269654 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlrlx" event={"ID":"88efab12-8605-40ee-9912-992901670b00","Type":"ContainerDied","Data":"237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf"} Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.269684 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlrlx" event={"ID":"88efab12-8605-40ee-9912-992901670b00","Type":"ContainerDied","Data":"04bd4ce9a132c460ce070bab4c10ae0e32c808371e6ca3328a05aaab53c2a696"} Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.269702 4966 scope.go:117] "RemoveContainer" containerID="237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.269851 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlrlx" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.332148 4966 scope.go:117] "RemoveContainer" containerID="ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.334656 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlrlx"] Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.337558 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xlrlx"] Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.354594 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bqqt\" (UniqueName: \"kubernetes.io/projected/0c2676b8-74c4-486a-b26a-d5dcc16e96dd-kube-api-access-2bqqt\") pod \"marketplace-operator-79b997595-grzxs\" (UID: \"0c2676b8-74c4-486a-b26a-d5dcc16e96dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.354629 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0c2676b8-74c4-486a-b26a-d5dcc16e96dd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-grzxs\" (UID: \"0c2676b8-74c4-486a-b26a-d5dcc16e96dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.354652 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0c2676b8-74c4-486a-b26a-d5dcc16e96dd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-grzxs\" (UID: \"0c2676b8-74c4-486a-b26a-d5dcc16e96dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.362645 4966 scope.go:117] "RemoveContainer" containerID="e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.373983 4966 scope.go:117] "RemoveContainer" containerID="237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.374321 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf\": container with ID starting with 237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf not found: ID does not exist" containerID="237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.374358 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf"} err="failed to get container status \"237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf\": rpc error: code = NotFound desc = could not find container \"237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf\": container with ID starting with 237284dd8c1fe8c2e394d1a9ff5bbb94ddf871543989b17a08cc9d7005aa72bf not found: ID does not exist" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.374383 4966 scope.go:117] "RemoveContainer" containerID="ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.374666 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427\": container with ID starting with ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427 not found: ID does not exist" containerID="ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.374706 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427"} err="failed to get container status \"ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427\": rpc error: code = NotFound desc = could not find container \"ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427\": container with ID starting with ab60b398cd9785e9a0d97589a2d577176e5c2997c072af57643bf6408af13427 not found: ID does not exist" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.374751 4966 scope.go:117] "RemoveContainer" containerID="e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6" Dec 17 08:25:21 crc kubenswrapper[4966]: E1217 08:25:21.374958 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6\": container with ID starting with e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6 not found: ID does not exist" containerID="e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.374981 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6"} err="failed to get container status \"e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6\": rpc error: code = NotFound desc = could not find container \"e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6\": container with ID starting with e0a58ca9f0d03da51d978d6afa42a2cd5c5a2ac9a2e54ae743059cecb4a747f6 not found: ID does not exist" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.455837 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bqqt\" (UniqueName: \"kubernetes.io/projected/0c2676b8-74c4-486a-b26a-d5dcc16e96dd-kube-api-access-2bqqt\") pod \"marketplace-operator-79b997595-grzxs\" (UID: \"0c2676b8-74c4-486a-b26a-d5dcc16e96dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.456152 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0c2676b8-74c4-486a-b26a-d5dcc16e96dd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-grzxs\" (UID: \"0c2676b8-74c4-486a-b26a-d5dcc16e96dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.456186 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0c2676b8-74c4-486a-b26a-d5dcc16e96dd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-grzxs\" (UID: \"0c2676b8-74c4-486a-b26a-d5dcc16e96dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.458280 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0c2676b8-74c4-486a-b26a-d5dcc16e96dd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-grzxs\" (UID: \"0c2676b8-74c4-486a-b26a-d5dcc16e96dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.463405 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0c2676b8-74c4-486a-b26a-d5dcc16e96dd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-grzxs\" (UID: \"0c2676b8-74c4-486a-b26a-d5dcc16e96dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.471628 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bqqt\" (UniqueName: \"kubernetes.io/projected/0c2676b8-74c4-486a-b26a-d5dcc16e96dd-kube-api-access-2bqqt\") pod \"marketplace-operator-79b997595-grzxs\" (UID: \"0c2676b8-74c4-486a-b26a-d5dcc16e96dd\") " pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.526449 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:21 crc kubenswrapper[4966]: I1217 08:25:21.922980 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-grzxs"] Dec 17 08:25:21 crc kubenswrapper[4966]: W1217 08:25:21.924314 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c2676b8_74c4_486a_b26a_d5dcc16e96dd.slice/crio-846979a71f51fdbda5954e5c12ae9d45800fe6eeb5b9a9bcf82e714cdd35e6b0 WatchSource:0}: Error finding container 846979a71f51fdbda5954e5c12ae9d45800fe6eeb5b9a9bcf82e714cdd35e6b0: Status 404 returned error can't find the container with id 846979a71f51fdbda5954e5c12ae9d45800fe6eeb5b9a9bcf82e714cdd35e6b0 Dec 17 08:25:22 crc kubenswrapper[4966]: E1217 08:25:22.201331 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8 is running failed: container process not found" containerID="c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8" cmd=["grpc_health_probe","-addr=:50051"] Dec 17 08:25:22 crc kubenswrapper[4966]: E1217 08:25:22.202046 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8 is running failed: container process not found" containerID="c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8" cmd=["grpc_health_probe","-addr=:50051"] Dec 17 08:25:22 crc kubenswrapper[4966]: E1217 08:25:22.202388 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8 is running failed: container process not found" containerID="c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8" cmd=["grpc_health_probe","-addr=:50051"] Dec 17 08:25:22 crc kubenswrapper[4966]: E1217 08:25:22.202459 4966 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-zvg6n" podUID="d5c75559-5110-464f-893d-421f55a5318f" containerName="registry-server" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.280618 4966 generic.go:334] "Generic (PLEG): container finished" podID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" containerID="03dd0839c60eb9aa19f2ad1e5c5977b922196b3dae99e97feaacb80b30c06d65" exitCode=0 Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.280689 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" event={"ID":"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c","Type":"ContainerDied","Data":"03dd0839c60eb9aa19f2ad1e5c5977b922196b3dae99e97feaacb80b30c06d65"} Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.281776 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" event={"ID":"0c2676b8-74c4-486a-b26a-d5dcc16e96dd","Type":"ContainerStarted","Data":"846979a71f51fdbda5954e5c12ae9d45800fe6eeb5b9a9bcf82e714cdd35e6b0"} Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.285433 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g57dd" event={"ID":"763b5661-27a9-4a61-9de9-17b7f8832720","Type":"ContainerDied","Data":"12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3"} Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.285491 4966 generic.go:334] "Generic (PLEG): container finished" podID="763b5661-27a9-4a61-9de9-17b7f8832720" containerID="12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3" exitCode=0 Dec 17 08:25:22 crc kubenswrapper[4966]: E1217 08:25:22.418102 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3 is running failed: container process not found" containerID="12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3" cmd=["grpc_health_probe","-addr=:50051"] Dec 17 08:25:22 crc kubenswrapper[4966]: E1217 08:25:22.418610 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3 is running failed: container process not found" containerID="12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3" cmd=["grpc_health_probe","-addr=:50051"] Dec 17 08:25:22 crc kubenswrapper[4966]: E1217 08:25:22.418903 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3 is running failed: container process not found" containerID="12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3" cmd=["grpc_health_probe","-addr=:50051"] Dec 17 08:25:22 crc kubenswrapper[4966]: E1217 08:25:22.418931 4966 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-g57dd" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" containerName="registry-server" Dec 17 08:25:22 crc kubenswrapper[4966]: E1217 08:25:22.502503 4966 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.688440 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.793430 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.837366 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88efab12-8605-40ee-9912-992901670b00" path="/var/lib/kubelet/pods/88efab12-8605-40ee-9912-992901670b00/volumes" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.869940 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.870220 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjxmn\" (UniqueName: \"kubernetes.io/projected/763b5661-27a9-4a61-9de9-17b7f8832720-kube-api-access-zjxmn\") pod \"763b5661-27a9-4a61-9de9-17b7f8832720\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.870263 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-catalog-content\") pod \"763b5661-27a9-4a61-9de9-17b7f8832720\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.870313 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-utilities\") pod \"763b5661-27a9-4a61-9de9-17b7f8832720\" (UID: \"763b5661-27a9-4a61-9de9-17b7f8832720\") " Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.871518 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-utilities" (OuterVolumeSpecName: "utilities") pod "763b5661-27a9-4a61-9de9-17b7f8832720" (UID: "763b5661-27a9-4a61-9de9-17b7f8832720"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.876407 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/763b5661-27a9-4a61-9de9-17b7f8832720-kube-api-access-zjxmn" (OuterVolumeSpecName: "kube-api-access-zjxmn") pod "763b5661-27a9-4a61-9de9-17b7f8832720" (UID: "763b5661-27a9-4a61-9de9-17b7f8832720"). InnerVolumeSpecName "kube-api-access-zjxmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.919689 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.959095 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "763b5661-27a9-4a61-9de9-17b7f8832720" (UID: "763b5661-27a9-4a61-9de9-17b7f8832720"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.971628 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-trusted-ca\") pod \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.971713 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-976x6\" (UniqueName: \"kubernetes.io/projected/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-kube-api-access-976x6\") pod \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.971737 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-operator-metrics\") pod \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\" (UID: \"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c\") " Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.971774 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-utilities\") pod \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.971838 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8dpr\" (UniqueName: \"kubernetes.io/projected/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-kube-api-access-n8dpr\") pod \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.972296 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-catalog-content\") pod \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\" (UID: \"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2\") " Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.972516 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-utilities" (OuterVolumeSpecName: "utilities") pod "c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" (UID: "c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.972660 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjxmn\" (UniqueName: \"kubernetes.io/projected/763b5661-27a9-4a61-9de9-17b7f8832720-kube-api-access-zjxmn\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.972672 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.972683 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763b5661-27a9-4a61-9de9-17b7f8832720-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.973234 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" (UID: "c5b35d14-4d2f-45a2-9c20-7aca870b0f5c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.974630 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-kube-api-access-n8dpr" (OuterVolumeSpecName: "kube-api-access-n8dpr") pod "c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" (UID: "c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2"). InnerVolumeSpecName "kube-api-access-n8dpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.975423 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-kube-api-access-976x6" (OuterVolumeSpecName: "kube-api-access-976x6") pod "c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" (UID: "c5b35d14-4d2f-45a2-9c20-7aca870b0f5c"). InnerVolumeSpecName "kube-api-access-976x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:22 crc kubenswrapper[4966]: I1217 08:25:22.977150 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" (UID: "c5b35d14-4d2f-45a2-9c20-7aca870b0f5c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.073924 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-catalog-content\") pod \"c6569dae-8fd2-4536-a645-51eed6946e41\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.074252 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8jh9\" (UniqueName: \"kubernetes.io/projected/c6569dae-8fd2-4536-a645-51eed6946e41-kube-api-access-g8jh9\") pod \"c6569dae-8fd2-4536-a645-51eed6946e41\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.074379 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-utilities\") pod \"c6569dae-8fd2-4536-a645-51eed6946e41\" (UID: \"c6569dae-8fd2-4536-a645-51eed6946e41\") " Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.074699 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8dpr\" (UniqueName: \"kubernetes.io/projected/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-kube-api-access-n8dpr\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.074803 4966 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.074916 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-976x6\" (UniqueName: \"kubernetes.io/projected/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-kube-api-access-976x6\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.075003 4966 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.075075 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.075381 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-utilities" (OuterVolumeSpecName: "utilities") pod "c6569dae-8fd2-4536-a645-51eed6946e41" (UID: "c6569dae-8fd2-4536-a645-51eed6946e41"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.077083 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6569dae-8fd2-4536-a645-51eed6946e41-kube-api-access-g8jh9" (OuterVolumeSpecName: "kube-api-access-g8jh9") pod "c6569dae-8fd2-4536-a645-51eed6946e41" (UID: "c6569dae-8fd2-4536-a645-51eed6946e41"). InnerVolumeSpecName "kube-api-access-g8jh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.093459 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6569dae-8fd2-4536-a645-51eed6946e41" (UID: "c6569dae-8fd2-4536-a645-51eed6946e41"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.105259 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" (UID: "c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.175826 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.175862 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8jh9\" (UniqueName: \"kubernetes.io/projected/c6569dae-8fd2-4536-a645-51eed6946e41-kube-api-access-g8jh9\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.175889 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6569dae-8fd2-4536-a645-51eed6946e41-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.175900 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.292594 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.292625 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-km2tl" event={"ID":"c5b35d14-4d2f-45a2-9c20-7aca870b0f5c","Type":"ContainerDied","Data":"9b7e44bfc85d5c569d47566faf974889bc46e58a3d5001bac59873c10aef7ac8"} Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.292701 4966 scope.go:117] "RemoveContainer" containerID="03dd0839c60eb9aa19f2ad1e5c5977b922196b3dae99e97feaacb80b30c06d65" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.296173 4966 generic.go:334] "Generic (PLEG): container finished" podID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerID="ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168" exitCode=0 Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.296224 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ssxxq" event={"ID":"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2","Type":"ContainerDied","Data":"ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168"} Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.296248 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ssxxq" event={"ID":"c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2","Type":"ContainerDied","Data":"ed993ff61ed1ac93f202bec5556728b8bfcb8a49285e1828217a394fd8957d56"} Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.296325 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ssxxq" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.325209 4966 scope.go:117] "RemoveContainer" containerID="ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.325959 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" event={"ID":"0c2676b8-74c4-486a-b26a-d5dcc16e96dd","Type":"ContainerStarted","Data":"6fcae1ee1e4c3d9da3e3e03d773cc31b416fac64871639086ba299f7a4014d2d"} Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.328859 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.330520 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.336084 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ssxxq"] Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.340115 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ssxxq"] Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.342056 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.345089 4966 generic.go:334] "Generic (PLEG): container finished" podID="d5c75559-5110-464f-893d-421f55a5318f" containerID="c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8" exitCode=0 Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.345250 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvg6n" event={"ID":"d5c75559-5110-464f-893d-421f55a5318f","Type":"ContainerDied","Data":"c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8"} Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.354697 4966 scope.go:117] "RemoveContainer" containerID="b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.356453 4966 generic.go:334] "Generic (PLEG): container finished" podID="c6569dae-8fd2-4536-a645-51eed6946e41" containerID="7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a" exitCode=0 Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.356522 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gllj5" event={"ID":"c6569dae-8fd2-4536-a645-51eed6946e41","Type":"ContainerDied","Data":"7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a"} Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.356550 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gllj5" event={"ID":"c6569dae-8fd2-4536-a645-51eed6946e41","Type":"ContainerDied","Data":"2dbcaad8e29b9b345684e76bacc17e8b225a3b469d5d0afddcea4b177c3befb6"} Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.356765 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gllj5" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.359600 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g57dd" event={"ID":"763b5661-27a9-4a61-9de9-17b7f8832720","Type":"ContainerDied","Data":"8c394aa8bee6f039909efc58bc38da4d9a5889634e46a4a3707eb3923db04ec7"} Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.359650 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g57dd" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.377666 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-grzxs" podStartSLOduration=2.377647983 podStartE2EDuration="2.377647983s" podCreationTimestamp="2025-12-17 08:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:25:23.354967593 +0000 UTC m=+258.900037535" watchObservedRunningTime="2025-12-17 08:25:23.377647983 +0000 UTC m=+258.922717925" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.390457 4966 scope.go:117] "RemoveContainer" containerID="d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.414264 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-km2tl"] Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.415595 4966 scope.go:117] "RemoveContainer" containerID="ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168" Dec 17 08:25:23 crc kubenswrapper[4966]: E1217 08:25:23.416071 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168\": container with ID starting with ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168 not found: ID does not exist" containerID="ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.416144 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168"} err="failed to get container status \"ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168\": rpc error: code = NotFound desc = could not find container \"ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168\": container with ID starting with ba5267fc477783a8c548f522a8d7aa6a122128185ac9e4c0e818c2942f4c7168 not found: ID does not exist" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.416172 4966 scope.go:117] "RemoveContainer" containerID="b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9" Dec 17 08:25:23 crc kubenswrapper[4966]: E1217 08:25:23.416529 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9\": container with ID starting with b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9 not found: ID does not exist" containerID="b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.416561 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9"} err="failed to get container status \"b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9\": rpc error: code = NotFound desc = could not find container \"b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9\": container with ID starting with b88acc983ceff0d72fca01d4727693ad755cd2eacb3ae666a79e57387483baf9 not found: ID does not exist" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.416581 4966 scope.go:117] "RemoveContainer" containerID="d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c" Dec 17 08:25:23 crc kubenswrapper[4966]: E1217 08:25:23.416770 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c\": container with ID starting with d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c not found: ID does not exist" containerID="d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.416794 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c"} err="failed to get container status \"d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c\": rpc error: code = NotFound desc = could not find container \"d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c\": container with ID starting with d41f5f3bc8fd1a9da44c0f6456a78f84cfb1d4990c2d9b86b27b3c53ebc7619c not found: ID does not exist" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.416811 4966 scope.go:117] "RemoveContainer" containerID="c985cdc722fec615598b46832c89083317e63ab8abdbafcdd6602a9b6e53cee8" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.416960 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-km2tl"] Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.427673 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gllj5"] Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.432434 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gllj5"] Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.440823 4966 scope.go:117] "RemoveContainer" containerID="52c0d1eb699e068249ad478405cc98ae555299fd794c4128ad677584b5acfa58" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.452300 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g57dd"] Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.454689 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g57dd"] Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.463618 4966 scope.go:117] "RemoveContainer" containerID="5efbceefa6eeb8f08adae2d3981dcf5e6ffdf658883bf68aa64219e10f229b2d" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.484830 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-utilities\") pod \"d5c75559-5110-464f-893d-421f55a5318f\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.485139 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psjr8\" (UniqueName: \"kubernetes.io/projected/d5c75559-5110-464f-893d-421f55a5318f-kube-api-access-psjr8\") pod \"d5c75559-5110-464f-893d-421f55a5318f\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.485183 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-catalog-content\") pod \"d5c75559-5110-464f-893d-421f55a5318f\" (UID: \"d5c75559-5110-464f-893d-421f55a5318f\") " Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.486252 4966 scope.go:117] "RemoveContainer" containerID="7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.486847 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-utilities" (OuterVolumeSpecName: "utilities") pod "d5c75559-5110-464f-893d-421f55a5318f" (UID: "d5c75559-5110-464f-893d-421f55a5318f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.491680 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c75559-5110-464f-893d-421f55a5318f-kube-api-access-psjr8" (OuterVolumeSpecName: "kube-api-access-psjr8") pod "d5c75559-5110-464f-893d-421f55a5318f" (UID: "d5c75559-5110-464f-893d-421f55a5318f"). InnerVolumeSpecName "kube-api-access-psjr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.501730 4966 scope.go:117] "RemoveContainer" containerID="f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.515731 4966 scope.go:117] "RemoveContainer" containerID="c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.529195 4966 scope.go:117] "RemoveContainer" containerID="7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a" Dec 17 08:25:23 crc kubenswrapper[4966]: E1217 08:25:23.529522 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a\": container with ID starting with 7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a not found: ID does not exist" containerID="7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.529551 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a"} err="failed to get container status \"7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a\": rpc error: code = NotFound desc = could not find container \"7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a\": container with ID starting with 7d93547b5b3650cec1f7a383ea6a9a9af87f72c6faa1705b827f2b79fc186a7a not found: ID does not exist" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.529573 4966 scope.go:117] "RemoveContainer" containerID="f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432" Dec 17 08:25:23 crc kubenswrapper[4966]: E1217 08:25:23.529786 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432\": container with ID starting with f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432 not found: ID does not exist" containerID="f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.529807 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432"} err="failed to get container status \"f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432\": rpc error: code = NotFound desc = could not find container \"f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432\": container with ID starting with f6688a91f819ea6d5f2e80973b858520e9cad9ae617ccd10bd5ff20cf97ba432 not found: ID does not exist" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.529820 4966 scope.go:117] "RemoveContainer" containerID="c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943" Dec 17 08:25:23 crc kubenswrapper[4966]: E1217 08:25:23.530427 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943\": container with ID starting with c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943 not found: ID does not exist" containerID="c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.530447 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943"} err="failed to get container status \"c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943\": rpc error: code = NotFound desc = could not find container \"c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943\": container with ID starting with c91fe53d0cfeac17f345dee1213bee4cc5d08b7f63438cf159aa78fc17ffc943 not found: ID does not exist" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.530460 4966 scope.go:117] "RemoveContainer" containerID="12914c4280f12500e073600fd81c15bede774bfe274852212c801e4ff400a0a3" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.535854 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5c75559-5110-464f-893d-421f55a5318f" (UID: "d5c75559-5110-464f-893d-421f55a5318f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.544216 4966 scope.go:117] "RemoveContainer" containerID="233bf7829ec27fc201d8c721d7ac53acf2b3f927e348fd828b6f99d16d99c949" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.556504 4966 scope.go:117] "RemoveContainer" containerID="21001dacd7269f5387581569ad4f2b486efd54f2c8bf3ea69bea0b53c1b716b8" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.586359 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.586389 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psjr8\" (UniqueName: \"kubernetes.io/projected/d5c75559-5110-464f-893d-421f55a5318f-kube-api-access-psjr8\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:23 crc kubenswrapper[4966]: I1217 08:25:23.586400 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5c75559-5110-464f-893d-421f55a5318f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112038 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qxgxs"] Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112256 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerName="extract-content" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112272 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerName="extract-content" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112286 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" containerName="extract-content" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112294 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" containerName="extract-content" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112308 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c75559-5110-464f-893d-421f55a5318f" containerName="extract-content" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112316 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c75559-5110-464f-893d-421f55a5318f" containerName="extract-content" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112326 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerName="extract-utilities" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112334 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerName="extract-utilities" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112346 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112354 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112362 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" containerName="extract-utilities" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112370 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" containerName="extract-utilities" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112380 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" containerName="extract-utilities" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112387 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" containerName="extract-utilities" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112395 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112401 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112410 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112417 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112429 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c75559-5110-464f-893d-421f55a5318f" containerName="extract-utilities" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112438 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c75559-5110-464f-893d-421f55a5318f" containerName="extract-utilities" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112446 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" containerName="extract-content" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112453 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" containerName="extract-content" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112462 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c75559-5110-464f-893d-421f55a5318f" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112470 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c75559-5110-464f-893d-421f55a5318f" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.112480 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" containerName="marketplace-operator" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112487 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" containerName="marketplace-operator" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112584 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c75559-5110-464f-893d-421f55a5318f" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112597 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112607 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" containerName="marketplace-operator" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112622 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.112634 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" containerName="registry-server" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.113424 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.115765 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.175062 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxgxs"] Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.194552 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90e381b2-c948-4b5a-b3e3-32210541dda2-catalog-content\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.194721 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90e381b2-c948-4b5a-b3e3-32210541dda2-utilities\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.194941 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nclvj\" (UniqueName: \"kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.205029 4966 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.205993 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.206868 4966 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.207782 4966 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208092 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc" gracePeriod=15 Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208147 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764" gracePeriod=15 Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208153 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf" gracePeriod=15 Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.208231 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208246 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.208257 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208204 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b" gracePeriod=15 Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208357 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732" gracePeriod=15 Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208267 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.208402 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208415 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.208428 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208437 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.208456 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208464 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.208492 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208506 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208722 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208740 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208752 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208766 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208778 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.208906 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.208916 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.209029 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.296118 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.296220 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclvj\" (UniqueName: \"kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.296243 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.296899 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.296932 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.296966 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.296997 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.296996 4966 projected.go:194] Error preparing data for projected volume kube-api-access-nclvj for pod openshift-marketplace/redhat-marketplace-qxgxs: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.297042 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90e381b2-c948-4b5a-b3e3-32210541dda2-catalog-content\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.297056 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj podName:90e381b2-c948-4b5a-b3e3-32210541dda2 nodeName:}" failed. No retries permitted until 2025-12-17 08:25:24.797038673 +0000 UTC m=+260.342108615 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-nclvj" (UniqueName: "kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj") pod "redhat-marketplace-qxgxs" (UID: "90e381b2-c948-4b5a-b3e3-32210541dda2") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.297400 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.297501 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.297461 4966 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-qxgxs.1881f331d74cdfd1 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-qxgxs,UID:90e381b2-c948-4b5a-b3e3-32210541dda2,APIVersion:v1,ResourceVersion:29552,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-nclvj\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token\": dial tcp 38.102.83.146:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-17 08:25:24.297031633 +0000 UTC m=+259.842101575,LastTimestamp:2025-12-17 08:25:24.297031633 +0000 UTC m=+259.842101575,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.297550 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90e381b2-c948-4b5a-b3e3-32210541dda2-utilities\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.297657 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90e381b2-c948-4b5a-b3e3-32210541dda2-catalog-content\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.297900 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90e381b2-c948-4b5a-b3e3-32210541dda2-utilities\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.367541 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvg6n" event={"ID":"d5c75559-5110-464f-893d-421f55a5318f","Type":"ContainerDied","Data":"13cdb043859552b8d74247cd5f8391eb672839fd8c05fe514e8517a5726cbb69"} Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.367638 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zvg6n" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.369325 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.369587 4966 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.383956 4966 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.384216 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.384704 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.386205 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.387110 4966 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc" exitCode=0 Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.387140 4966 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf" exitCode=0 Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.387150 4966 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764" exitCode=0 Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.387160 4966 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b" exitCode=2 Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.387166 4966 scope.go:117] "RemoveContainer" containerID="101f72456f3eb0763ec4057064b166b2659e1e1ffd388586c43c2337d8b6adb7" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.398722 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.398803 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.398822 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.398840 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.398912 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.398948 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.398920 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.398972 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.399038 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.399096 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.399137 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.399154 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.399170 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.399200 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.399206 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.399299 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.805316 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclvj\" (UniqueName: \"kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.805993 4966 projected.go:194] Error preparing data for projected volume kube-api-access-nclvj for pod openshift-marketplace/redhat-marketplace-qxgxs: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:24 crc kubenswrapper[4966]: E1217 08:25:24.806053 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj podName:90e381b2-c948-4b5a-b3e3-32210541dda2 nodeName:}" failed. No retries permitted until 2025-12-17 08:25:25.806036804 +0000 UTC m=+261.351106756 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-nclvj" (UniqueName: "kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj") pod "redhat-marketplace-qxgxs" (UID: "90e381b2-c948-4b5a-b3e3-32210541dda2") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.835581 4966 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.835920 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.843225 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="763b5661-27a9-4a61-9de9-17b7f8832720" path="/var/lib/kubelet/pods/763b5661-27a9-4a61-9de9-17b7f8832720/volumes" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.843868 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5b35d14-4d2f-45a2-9c20-7aca870b0f5c" path="/var/lib/kubelet/pods/c5b35d14-4d2f-45a2-9c20-7aca870b0f5c/volumes" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.844301 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6569dae-8fd2-4536-a645-51eed6946e41" path="/var/lib/kubelet/pods/c6569dae-8fd2-4536-a645-51eed6946e41/volumes" Dec 17 08:25:24 crc kubenswrapper[4966]: I1217 08:25:24.845288 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2" path="/var/lib/kubelet/pods/c82ff2ce-39a3-48bb-a4d3-69cfe999b8a2/volumes" Dec 17 08:25:25 crc kubenswrapper[4966]: E1217 08:25:25.145976 4966 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-qxgxs.1881f331d74cdfd1 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-qxgxs,UID:90e381b2-c948-4b5a-b3e3-32210541dda2,APIVersion:v1,ResourceVersion:29552,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-nclvj\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token\": dial tcp 38.102.83.146:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-17 08:25:24.297031633 +0000 UTC m=+259.842101575,LastTimestamp:2025-12-17 08:25:24.297031633 +0000 UTC m=+259.842101575,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 17 08:25:25 crc kubenswrapper[4966]: I1217 08:25:25.398221 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 17 08:25:25 crc kubenswrapper[4966]: I1217 08:25:25.401574 4966 generic.go:334] "Generic (PLEG): container finished" podID="771b39e4-c3a9-4363-b57f-a429cc16157e" containerID="297409173e166c1e3d780663674d32edbb3e402d8f37354a8d54bb02bdcc01bb" exitCode=0 Dec 17 08:25:25 crc kubenswrapper[4966]: I1217 08:25:25.401636 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"771b39e4-c3a9-4363-b57f-a429cc16157e","Type":"ContainerDied","Data":"297409173e166c1e3d780663674d32edbb3e402d8f37354a8d54bb02bdcc01bb"} Dec 17 08:25:25 crc kubenswrapper[4966]: I1217 08:25:25.402545 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:25 crc kubenswrapper[4966]: I1217 08:25:25.402914 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:25 crc kubenswrapper[4966]: I1217 08:25:25.820792 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclvj\" (UniqueName: \"kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:25 crc kubenswrapper[4966]: E1217 08:25:25.821799 4966 projected.go:194] Error preparing data for projected volume kube-api-access-nclvj for pod openshift-marketplace/redhat-marketplace-qxgxs: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:25 crc kubenswrapper[4966]: E1217 08:25:25.821932 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj podName:90e381b2-c948-4b5a-b3e3-32210541dda2 nodeName:}" failed. No retries permitted until 2025-12-17 08:25:27.821863046 +0000 UTC m=+263.366933018 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-nclvj" (UniqueName: "kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj") pod "redhat-marketplace-qxgxs" (UID: "90e381b2-c948-4b5a-b3e3-32210541dda2") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.598067 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.600651 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.601190 4966 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.601351 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.601608 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.630337 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.630399 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.630445 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.630487 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.630520 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.630536 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.630758 4966 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.630772 4966 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.630780 4966 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.636767 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.637292 4966 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.637572 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.637773 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.731229 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/771b39e4-c3a9-4363-b57f-a429cc16157e-kube-api-access\") pod \"771b39e4-c3a9-4363-b57f-a429cc16157e\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.731282 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-var-lock\") pod \"771b39e4-c3a9-4363-b57f-a429cc16157e\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.731322 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-kubelet-dir\") pod \"771b39e4-c3a9-4363-b57f-a429cc16157e\" (UID: \"771b39e4-c3a9-4363-b57f-a429cc16157e\") " Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.731425 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-var-lock" (OuterVolumeSpecName: "var-lock") pod "771b39e4-c3a9-4363-b57f-a429cc16157e" (UID: "771b39e4-c3a9-4363-b57f-a429cc16157e"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.731691 4966 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-var-lock\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.731714 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "771b39e4-c3a9-4363-b57f-a429cc16157e" (UID: "771b39e4-c3a9-4363-b57f-a429cc16157e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.748583 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/771b39e4-c3a9-4363-b57f-a429cc16157e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "771b39e4-c3a9-4363-b57f-a429cc16157e" (UID: "771b39e4-c3a9-4363-b57f-a429cc16157e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.832205 4966 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/771b39e4-c3a9-4363-b57f-a429cc16157e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.832238 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/771b39e4-c3a9-4363-b57f-a429cc16157e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:26 crc kubenswrapper[4966]: I1217 08:25:26.837042 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.413693 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.415155 4966 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732" exitCode=0 Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.415271 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.415278 4966 scope.go:117] "RemoveContainer" containerID="5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.416740 4966 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.418593 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.418884 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.419452 4966 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.419619 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.419772 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.428490 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"771b39e4-c3a9-4363-b57f-a429cc16157e","Type":"ContainerDied","Data":"31ce221e8475f59b458301dd2cd6d659f57d1d265ccac64766e52be50e3bd317"} Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.428520 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31ce221e8475f59b458301dd2cd6d659f57d1d265ccac64766e52be50e3bd317" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.428572 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.433650 4966 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.435676 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.436141 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.447743 4966 scope.go:117] "RemoveContainer" containerID="5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.462346 4966 scope.go:117] "RemoveContainer" containerID="5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.486276 4966 scope.go:117] "RemoveContainer" containerID="bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.506110 4966 scope.go:117] "RemoveContainer" containerID="58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.527539 4966 scope.go:117] "RemoveContainer" containerID="a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.562369 4966 scope.go:117] "RemoveContainer" containerID="5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc" Dec 17 08:25:27 crc kubenswrapper[4966]: E1217 08:25:27.562960 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\": container with ID starting with 5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc not found: ID does not exist" containerID="5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.563012 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc"} err="failed to get container status \"5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\": rpc error: code = NotFound desc = could not find container \"5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc\": container with ID starting with 5dd0dd28c8783eb2278161d1f72b468d01784e3b7acc0561b79d7409942fa6dc not found: ID does not exist" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.563062 4966 scope.go:117] "RemoveContainer" containerID="5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf" Dec 17 08:25:27 crc kubenswrapper[4966]: E1217 08:25:27.564096 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\": container with ID starting with 5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf not found: ID does not exist" containerID="5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.564137 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf"} err="failed to get container status \"5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\": rpc error: code = NotFound desc = could not find container \"5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf\": container with ID starting with 5a738e1d2993f64d3676f0b95e8717e2ded8ad50f60e68dbecd6e6a4e8a802cf not found: ID does not exist" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.564183 4966 scope.go:117] "RemoveContainer" containerID="5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764" Dec 17 08:25:27 crc kubenswrapper[4966]: E1217 08:25:27.564833 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\": container with ID starting with 5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764 not found: ID does not exist" containerID="5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.564861 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764"} err="failed to get container status \"5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\": rpc error: code = NotFound desc = could not find container \"5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764\": container with ID starting with 5132f1aaece1a77d3a701072ad8a5aa40cdb5ba5a431d20e5878d4aaf1fb5764 not found: ID does not exist" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.564903 4966 scope.go:117] "RemoveContainer" containerID="bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b" Dec 17 08:25:27 crc kubenswrapper[4966]: E1217 08:25:27.565245 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\": container with ID starting with bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b not found: ID does not exist" containerID="bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.565295 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b"} err="failed to get container status \"bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\": rpc error: code = NotFound desc = could not find container \"bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b\": container with ID starting with bb8e84cddf9f64881bd0ed6a90df78f6ce7406dbaf77840ff57eda7e8f57e42b not found: ID does not exist" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.565324 4966 scope.go:117] "RemoveContainer" containerID="58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732" Dec 17 08:25:27 crc kubenswrapper[4966]: E1217 08:25:27.565729 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\": container with ID starting with 58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732 not found: ID does not exist" containerID="58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.565800 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732"} err="failed to get container status \"58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\": rpc error: code = NotFound desc = could not find container \"58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732\": container with ID starting with 58a1f66199ad1edc9c67a42e5b9df3bc7be1e3edfbeaec8678a965111e543732 not found: ID does not exist" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.565847 4966 scope.go:117] "RemoveContainer" containerID="a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466" Dec 17 08:25:27 crc kubenswrapper[4966]: E1217 08:25:27.566215 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\": container with ID starting with a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466 not found: ID does not exist" containerID="a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.566258 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466"} err="failed to get container status \"a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\": rpc error: code = NotFound desc = could not find container \"a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466\": container with ID starting with a4b1013e6f45837027339dc9214c4753e6b9243a3335f7eb45ba12093ffad466 not found: ID does not exist" Dec 17 08:25:27 crc kubenswrapper[4966]: I1217 08:25:27.845169 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclvj\" (UniqueName: \"kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:27 crc kubenswrapper[4966]: E1217 08:25:27.845572 4966 projected.go:194] Error preparing data for projected volume kube-api-access-nclvj for pod openshift-marketplace/redhat-marketplace-qxgxs: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:27 crc kubenswrapper[4966]: E1217 08:25:27.845625 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj podName:90e381b2-c948-4b5a-b3e3-32210541dda2 nodeName:}" failed. No retries permitted until 2025-12-17 08:25:31.845608037 +0000 UTC m=+267.390677979 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-nclvj" (UniqueName: "kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj") pod "redhat-marketplace-qxgxs" (UID: "90e381b2-c948-4b5a-b3e3-32210541dda2") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:29 crc kubenswrapper[4966]: E1217 08:25:29.239367 4966 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:29 crc kubenswrapper[4966]: I1217 08:25:29.239994 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:29 crc kubenswrapper[4966]: I1217 08:25:29.442343 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2db4250d82dde4e302d6ad4faad836cfe8d3fb6eb3ebb27ac2b5f4b3d3452456"} Dec 17 08:25:30 crc kubenswrapper[4966]: E1217 08:25:30.195829 4966 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:30 crc kubenswrapper[4966]: E1217 08:25:30.196794 4966 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:30 crc kubenswrapper[4966]: E1217 08:25:30.197253 4966 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:30 crc kubenswrapper[4966]: E1217 08:25:30.197653 4966 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:30 crc kubenswrapper[4966]: E1217 08:25:30.198445 4966 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:30 crc kubenswrapper[4966]: I1217 08:25:30.198482 4966 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 17 08:25:30 crc kubenswrapper[4966]: E1217 08:25:30.198654 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="200ms" Dec 17 08:25:30 crc kubenswrapper[4966]: E1217 08:25:30.399321 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="400ms" Dec 17 08:25:30 crc kubenswrapper[4966]: I1217 08:25:30.448391 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"01f9f9b46c93b0fd3ed051c0452cdf4fdb69f5814410d5dddb6802b2ab199b90"} Dec 17 08:25:30 crc kubenswrapper[4966]: E1217 08:25:30.448962 4966 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:30 crc kubenswrapper[4966]: I1217 08:25:30.448977 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:30 crc kubenswrapper[4966]: I1217 08:25:30.449204 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:30 crc kubenswrapper[4966]: E1217 08:25:30.801061 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="800ms" Dec 17 08:25:31 crc kubenswrapper[4966]: E1217 08:25:31.454479 4966 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:25:31 crc kubenswrapper[4966]: E1217 08:25:31.602470 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="1.6s" Dec 17 08:25:31 crc kubenswrapper[4966]: I1217 08:25:31.899839 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclvj\" (UniqueName: \"kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:31 crc kubenswrapper[4966]: E1217 08:25:31.900622 4966 projected.go:194] Error preparing data for projected volume kube-api-access-nclvj for pod openshift-marketplace/redhat-marketplace-qxgxs: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:31 crc kubenswrapper[4966]: E1217 08:25:31.900715 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj podName:90e381b2-c948-4b5a-b3e3-32210541dda2 nodeName:}" failed. No retries permitted until 2025-12-17 08:25:39.900690717 +0000 UTC m=+275.445760669 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-nclvj" (UniqueName: "kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj") pod "redhat-marketplace-qxgxs" (UID: "90e381b2-c948-4b5a-b3e3-32210541dda2") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:32 crc kubenswrapper[4966]: E1217 08:25:32.715307 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:25:32Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:25:32Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:25:32Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-17T08:25:32Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:32 crc kubenswrapper[4966]: E1217 08:25:32.715917 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:32 crc kubenswrapper[4966]: E1217 08:25:32.716294 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:32 crc kubenswrapper[4966]: E1217 08:25:32.716763 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:32 crc kubenswrapper[4966]: E1217 08:25:32.717274 4966 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:32 crc kubenswrapper[4966]: E1217 08:25:32.717312 4966 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 17 08:25:32 crc kubenswrapper[4966]: I1217 08:25:32.812492 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:25:32 crc kubenswrapper[4966]: I1217 08:25:32.812571 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:25:32 crc kubenswrapper[4966]: I1217 08:25:32.812627 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:25:32 crc kubenswrapper[4966]: I1217 08:25:32.812701 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:25:32 crc kubenswrapper[4966]: W1217 08:25:32.813428 4966 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/configmaps?fieldSelector=metadata.name%3Dnetworking-console-plugin&resourceVersion=27182": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:32 crc kubenswrapper[4966]: E1217 08:25:32.813530 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/configmaps?fieldSelector=metadata.name%3Dnetworking-console-plugin&resourceVersion=27182\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:25:32 crc kubenswrapper[4966]: W1217 08:25:32.813610 4966 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin-cert": failed to list *v1.Secret: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27177": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:32 crc kubenswrapper[4966]: E1217 08:25:32.813762 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27177\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:25:32 crc kubenswrapper[4966]: W1217 08:25:32.813905 4966 reflector.go:561] object-"openshift-network-diagnostics"/"kube-root-ca.crt": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27182": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:32 crc kubenswrapper[4966]: E1217 08:25:32.814095 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27182\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:25:33 crc kubenswrapper[4966]: E1217 08:25:33.204000 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="3.2s" Dec 17 08:25:33 crc kubenswrapper[4966]: E1217 08:25:33.812779 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:33 crc kubenswrapper[4966]: E1217 08:25:33.812929 4966 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: failed to sync secret cache: timed out waiting for the condition Dec 17 08:25:33 crc kubenswrapper[4966]: E1217 08:25:33.813047 4966 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:33 crc kubenswrapper[4966]: E1217 08:25:33.813063 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:33 crc kubenswrapper[4966]: E1217 08:25:33.813015 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:27:35.812990389 +0000 UTC m=+391.358060341 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync secret cache: timed out waiting for the condition Dec 17 08:25:33 crc kubenswrapper[4966]: E1217 08:25:33.813127 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-17 08:27:35.813109132 +0000 UTC m=+391.358179094 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:33 crc kubenswrapper[4966]: W1217 08:25:33.813376 4966 reflector.go:561] object-"openshift-network-diagnostics"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&resourceVersion=27182": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:33 crc kubenswrapper[4966]: E1217 08:25:33.813466 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&resourceVersion=27182\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:25:33 crc kubenswrapper[4966]: E1217 08:25:33.912211 4966 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" volumeName="registry-storage" Dec 17 08:25:34 crc kubenswrapper[4966]: W1217 08:25:34.744965 4966 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin-cert": failed to list *v1.Secret: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27177": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.745018 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27177\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.813848 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.813902 4966 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.813945 4966 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.813993 4966 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.813970 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-17 08:27:36.813950818 +0000 UTC m=+392.359020760 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.814082 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-17 08:27:36.814069812 +0000 UTC m=+392.359139754 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : failed to sync configmap cache: timed out waiting for the condition Dec 17 08:25:34 crc kubenswrapper[4966]: I1217 08:25:34.840109 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:34 crc kubenswrapper[4966]: I1217 08:25:34.841670 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.844662 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[networking-console-plugin-cert nginx-conf], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.852574 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-cqllr], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.859800 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-s2dwl], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 17 08:25:34 crc kubenswrapper[4966]: W1217 08:25:34.877349 4966 reflector.go:561] object-"openshift-network-diagnostics"/"kube-root-ca.crt": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27182": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:34 crc kubenswrapper[4966]: E1217 08:25:34.877424 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27182\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:25:35 crc kubenswrapper[4966]: E1217 08:25:35.147836 4966 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-qxgxs.1881f331d74cdfd1 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-qxgxs,UID:90e381b2-c948-4b5a-b3e3-32210541dda2,APIVersion:v1,ResourceVersion:29552,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-nclvj\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-marketplace/token\": dial tcp 38.102.83.146:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-17 08:25:24.297031633 +0000 UTC m=+259.842101575,LastTimestamp:2025-12-17 08:25:24.297031633 +0000 UTC m=+259.842101575,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 17 08:25:35 crc kubenswrapper[4966]: W1217 08:25:35.442094 4966 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/configmaps?fieldSelector=metadata.name%3Dnetworking-console-plugin&resourceVersion=27182": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:35 crc kubenswrapper[4966]: E1217 08:25:35.442185 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/configmaps?fieldSelector=metadata.name%3Dnetworking-console-plugin&resourceVersion=27182\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:25:36 crc kubenswrapper[4966]: W1217 08:25:36.189096 4966 reflector.go:561] object-"openshift-network-diagnostics"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&resourceVersion=27182": dial tcp 38.102.83.146:6443: connect: connection refused Dec 17 08:25:36 crc kubenswrapper[4966]: E1217 08:25:36.189202 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&resourceVersion=27182\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Dec 17 08:25:36 crc kubenswrapper[4966]: E1217 08:25:36.405185 4966 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="6.4s" Dec 17 08:25:36 crc kubenswrapper[4966]: I1217 08:25:36.830235 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:36 crc kubenswrapper[4966]: I1217 08:25:36.831641 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:36 crc kubenswrapper[4966]: I1217 08:25:36.833053 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:36 crc kubenswrapper[4966]: I1217 08:25:36.857719 4966 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:36 crc kubenswrapper[4966]: I1217 08:25:36.857777 4966 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:36 crc kubenswrapper[4966]: E1217 08:25:36.858738 4966 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:36 crc kubenswrapper[4966]: I1217 08:25:36.859497 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.496751 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.497130 4966 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf" exitCode=1 Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.497176 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf"} Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.497908 4966 scope.go:117] "RemoveContainer" containerID="ad62607c27c575a120d1421195b78d4d2c73fbafd324eead42a95511a71a1baf" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.498423 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.499031 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.499435 4966 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.500527 4966 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="c9c0c76695837ce7950005f0eeb6c8858f9f3c829e8a19dc00aa7533b00e5291" exitCode=0 Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.500577 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"c9c0c76695837ce7950005f0eeb6c8858f9f3c829e8a19dc00aa7533b00e5291"} Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.500609 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a095e8b78abeb205be2aad4c0d201dc882d642443b3e2d0b2ebb190683adaa78"} Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.501033 4966 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.501075 4966 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.501454 4966 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:37 crc kubenswrapper[4966]: E1217 08:25:37.501506 4966 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.502156 4966 status_manager.go:851] "Failed to get status for pod" podUID="d5c75559-5110-464f-893d-421f55a5318f" pod="openshift-marketplace/community-operators-zvg6n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zvg6n\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:37 crc kubenswrapper[4966]: I1217 08:25:37.502804 4966 status_manager.go:851] "Failed to get status for pod" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Dec 17 08:25:38 crc kubenswrapper[4966]: I1217 08:25:38.468847 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:25:38 crc kubenswrapper[4966]: I1217 08:25:38.509007 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 17 08:25:38 crc kubenswrapper[4966]: I1217 08:25:38.509079 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f533ecefa4cac89268edc07260f1e280ef8da98bae16bea8aa295bf9b556da88"} Dec 17 08:25:38 crc kubenswrapper[4966]: I1217 08:25:38.512759 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3e224516f429d8b339e2a2b6171ce9aecef88e16e20fe776b62bbfaaf63418a8"} Dec 17 08:25:38 crc kubenswrapper[4966]: I1217 08:25:38.512810 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"54343b1949276424e97941024cda027e92836d7bb4a8ce8c7c5b08b48d2f4e3b"} Dec 17 08:25:38 crc kubenswrapper[4966]: I1217 08:25:38.512820 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c8c6b5318771ec042ebc81ae503ad23eac57d0f79b8b7cb7e8a4646b29fa5c9d"} Dec 17 08:25:39 crc kubenswrapper[4966]: I1217 08:25:39.273912 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:25:39 crc kubenswrapper[4966]: I1217 08:25:39.285103 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:25:39 crc kubenswrapper[4966]: I1217 08:25:39.525530 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c433ebf8d209b95ec4b3b9f89b028cb844d989fdad2b5bbaee1526fcb4bf2077"} Dec 17 08:25:39 crc kubenswrapper[4966]: I1217 08:25:39.525931 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:25:39 crc kubenswrapper[4966]: I1217 08:25:39.525980 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"833e63e381125cf905709eb6bf5890c5bea9e60fcaf1df5cdd3180148a59fd6e"} Dec 17 08:25:39 crc kubenswrapper[4966]: I1217 08:25:39.525829 4966 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:39 crc kubenswrapper[4966]: I1217 08:25:39.526025 4966 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:39 crc kubenswrapper[4966]: I1217 08:25:39.901354 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclvj\" (UniqueName: \"kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:41 crc kubenswrapper[4966]: I1217 08:25:41.860795 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:41 crc kubenswrapper[4966]: I1217 08:25:41.860889 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:41 crc kubenswrapper[4966]: I1217 08:25:41.867436 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:43 crc kubenswrapper[4966]: I1217 08:25:43.923477 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" podUID="021d4c31-6de9-4e76-ac80-0f345b3d7f81" containerName="oauth-openshift" containerID="cri-o://bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9" gracePeriod=15 Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.393555 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.541973 4966 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.558152 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.568484 4966 generic.go:334] "Generic (PLEG): container finished" podID="021d4c31-6de9-4e76-ac80-0f345b3d7f81" containerID="bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9" exitCode=0 Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.568554 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.568591 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" event={"ID":"021d4c31-6de9-4e76-ac80-0f345b3d7f81","Type":"ContainerDied","Data":"bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9"} Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.568643 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nxfxg" event={"ID":"021d4c31-6de9-4e76-ac80-0f345b3d7f81","Type":"ContainerDied","Data":"72522b04e697435317f15de15d714dcd91f69c45b90c8733913d327b93bf61a6"} Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.568668 4966 scope.go:117] "RemoveContainer" containerID="bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.568958 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.569048 4966 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.569068 4966 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.573678 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583486 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-error\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583540 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-session\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583592 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-ocp-branding-template\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583622 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-trusted-ca-bundle\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583646 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-router-certs\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583704 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-policies\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583739 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-service-ca\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583770 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4n9j\" (UniqueName: \"kubernetes.io/projected/021d4c31-6de9-4e76-ac80-0f345b3d7f81-kube-api-access-q4n9j\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583799 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-login\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583841 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-idp-0-file-data\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583884 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-dir\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.583991 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-serving-cert\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.584031 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-cliconfig\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.584059 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-provider-selection\") pod \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\" (UID: \"021d4c31-6de9-4e76-ac80-0f345b3d7f81\") " Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.585222 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.585255 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.588472 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.589497 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.589680 4966 scope.go:117] "RemoveContainer" containerID="bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.589831 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.596066 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: E1217 08:25:44.596324 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9\": container with ID starting with bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9 not found: ID does not exist" containerID="bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.596361 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9"} err="failed to get container status \"bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9\": rpc error: code = NotFound desc = could not find container \"bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9\": container with ID starting with bd5e745576557c9c6dd532a92c5bc2d9f057685ed9cc90e3994420ab5f05f4f9 not found: ID does not exist" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.598302 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.603980 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.605753 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/021d4c31-6de9-4e76-ac80-0f345b3d7f81-kube-api-access-q4n9j" (OuterVolumeSpecName: "kube-api-access-q4n9j") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "kube-api-access-q4n9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.611833 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.614209 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.614601 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.614785 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.615007 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "021d4c31-6de9-4e76-ac80-0f345b3d7f81" (UID: "021d4c31-6de9-4e76-ac80-0f345b3d7f81"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685494 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685538 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685554 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685569 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685582 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685594 4966 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685606 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685620 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4n9j\" (UniqueName: \"kubernetes.io/projected/021d4c31-6de9-4e76-ac80-0f345b3d7f81-kube-api-access-q4n9j\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685634 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685645 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685656 4966 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/021d4c31-6de9-4e76-ac80-0f345b3d7f81-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685668 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685680 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.685692 4966 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/021d4c31-6de9-4e76-ac80-0f345b3d7f81-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.750807 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.843949 4966 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f6293698-8574-4d46-bc8b-3988101434d8" Dec 17 08:25:44 crc kubenswrapper[4966]: I1217 08:25:44.918856 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nclvj\" (UniqueName: \"kubernetes.io/projected/90e381b2-c948-4b5a-b3e3-32210541dda2-kube-api-access-nclvj\") pod \"redhat-marketplace-qxgxs\" (UID: \"90e381b2-c948-4b5a-b3e3-32210541dda2\") " pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:45 crc kubenswrapper[4966]: I1217 08:25:45.127397 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:45 crc kubenswrapper[4966]: I1217 08:25:45.182561 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 17 08:25:45 crc kubenswrapper[4966]: I1217 08:25:45.358952 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 17 08:25:45 crc kubenswrapper[4966]: I1217 08:25:45.576901 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxgxs" event={"ID":"90e381b2-c948-4b5a-b3e3-32210541dda2","Type":"ContainerStarted","Data":"2489b92a86bf69b405c8ec770e178fce36d31e19b7d51b3ec904c699cbcd0153"} Dec 17 08:25:45 crc kubenswrapper[4966]: I1217 08:25:45.577622 4966 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:45 crc kubenswrapper[4966]: I1217 08:25:45.577646 4966 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:45 crc kubenswrapper[4966]: I1217 08:25:45.580371 4966 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f6293698-8574-4d46-bc8b-3988101434d8" Dec 17 08:25:45 crc kubenswrapper[4966]: E1217 08:25:45.606625 4966 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"audit\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Dec 17 08:25:45 crc kubenswrapper[4966]: I1217 08:25:45.830183 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:25:46 crc kubenswrapper[4966]: I1217 08:25:46.585638 4966 generic.go:334] "Generic (PLEG): container finished" podID="90e381b2-c948-4b5a-b3e3-32210541dda2" containerID="3fb57b973278fdc1283d074bba50cb6a17a9ff8321087bea2a0b1ac7310fe6b2" exitCode=0 Dec 17 08:25:46 crc kubenswrapper[4966]: I1217 08:25:46.585689 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxgxs" event={"ID":"90e381b2-c948-4b5a-b3e3-32210541dda2","Type":"ContainerDied","Data":"3fb57b973278fdc1283d074bba50cb6a17a9ff8321087bea2a0b1ac7310fe6b2"} Dec 17 08:25:46 crc kubenswrapper[4966]: I1217 08:25:46.586313 4966 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:46 crc kubenswrapper[4966]: I1217 08:25:46.586327 4966 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:46 crc kubenswrapper[4966]: I1217 08:25:46.607716 4966 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f6293698-8574-4d46-bc8b-3988101434d8" Dec 17 08:25:46 crc kubenswrapper[4966]: I1217 08:25:46.830136 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:25:47 crc kubenswrapper[4966]: I1217 08:25:47.592911 4966 generic.go:334] "Generic (PLEG): container finished" podID="90e381b2-c948-4b5a-b3e3-32210541dda2" containerID="1a96a1bb12a2c1181711bee8ddcf46439d6afce6fc6e6ea884f52e2c8ece42ca" exitCode=0 Dec 17 08:25:47 crc kubenswrapper[4966]: I1217 08:25:47.593095 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxgxs" event={"ID":"90e381b2-c948-4b5a-b3e3-32210541dda2","Type":"ContainerDied","Data":"1a96a1bb12a2c1181711bee8ddcf46439d6afce6fc6e6ea884f52e2c8ece42ca"} Dec 17 08:25:48 crc kubenswrapper[4966]: I1217 08:25:48.473737 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 17 08:25:48 crc kubenswrapper[4966]: I1217 08:25:48.605836 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qxgxs" event={"ID":"90e381b2-c948-4b5a-b3e3-32210541dda2","Type":"ContainerStarted","Data":"0feda63e1fe6f505c0e605247587e3290d135b952ea7c25ad766d10964d80394"} Dec 17 08:25:49 crc kubenswrapper[4966]: I1217 08:25:49.830427 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:25:55 crc kubenswrapper[4966]: I1217 08:25:55.127697 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:55 crc kubenswrapper[4966]: I1217 08:25:55.128131 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:55 crc kubenswrapper[4966]: I1217 08:25:55.167510 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:55 crc kubenswrapper[4966]: I1217 08:25:55.416646 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 17 08:25:55 crc kubenswrapper[4966]: I1217 08:25:55.535175 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 17 08:25:55 crc kubenswrapper[4966]: I1217 08:25:55.545775 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 17 08:25:55 crc kubenswrapper[4966]: I1217 08:25:55.677726 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 17 08:25:55 crc kubenswrapper[4966]: I1217 08:25:55.698937 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qxgxs" Dec 17 08:25:55 crc kubenswrapper[4966]: I1217 08:25:55.892833 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 17 08:25:56 crc kubenswrapper[4966]: I1217 08:25:56.071089 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 17 08:25:56 crc kubenswrapper[4966]: I1217 08:25:56.745996 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 17 08:25:56 crc kubenswrapper[4966]: I1217 08:25:56.761227 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 17 08:25:56 crc kubenswrapper[4966]: I1217 08:25:56.895217 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 17 08:25:56 crc kubenswrapper[4966]: I1217 08:25:56.908574 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.030969 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.079215 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.081424 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.121661 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.124981 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.144015 4966 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.147087 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qxgxs" podStartSLOduration=31.426993161 podStartE2EDuration="33.147069259s" podCreationTimestamp="2025-12-17 08:25:24 +0000 UTC" firstStartedPulling="2025-12-17 08:25:46.589359022 +0000 UTC m=+282.134428964" lastFinishedPulling="2025-12-17 08:25:48.3094351 +0000 UTC m=+283.854505062" observedRunningTime="2025-12-17 08:25:48.624270867 +0000 UTC m=+284.169340819" watchObservedRunningTime="2025-12-17 08:25:57.147069259 +0000 UTC m=+292.692139211" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.152248 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-nxfxg","openshift-marketplace/community-operators-zvg6n"] Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.152317 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-6b788bb46c-9pxdc"] Dec 17 08:25:57 crc kubenswrapper[4966]: E1217 08:25:57.152483 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="021d4c31-6de9-4e76-ac80-0f345b3d7f81" containerName="oauth-openshift" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.152500 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="021d4c31-6de9-4e76-ac80-0f345b3d7f81" containerName="oauth-openshift" Dec 17 08:25:57 crc kubenswrapper[4966]: E1217 08:25:57.152514 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" containerName="installer" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.152523 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" containerName="installer" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.152748 4966 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.152846 4966 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="00ec6e41-e778-4f9e-98b8-593f1ec04a03" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.153082 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="021d4c31-6de9-4e76-ac80-0f345b3d7f81" containerName="oauth-openshift" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.153225 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="771b39e4-c3a9-4363-b57f-a429cc16157e" containerName="installer" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.153776 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qxgxs"] Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.153999 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.159169 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.159465 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.159802 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.159975 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.160137 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.162561 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.163049 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.163267 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.163543 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.163609 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.163717 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.163767 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.163275 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.171697 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.179091 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.188699 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.208197 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=13.208175986 podStartE2EDuration="13.208175986s" podCreationTimestamp="2025-12-17 08:25:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:25:57.203374493 +0000 UTC m=+292.748444455" watchObservedRunningTime="2025-12-17 08:25:57.208175986 +0000 UTC m=+292.753245948" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.321779 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.353363 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.353590 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-template-error\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.353711 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.353801 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f71a6b20-21c9-4920-a510-a5817462836a-audit-dir\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.353947 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.354074 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-template-login\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.354196 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-audit-policies\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.354452 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.354548 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.354589 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.354660 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.354694 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.354736 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-session\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.354764 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7745\" (UniqueName: \"kubernetes.io/projected/f71a6b20-21c9-4920-a510-a5817462836a-kube-api-access-v7745\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.363750 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.455547 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.455636 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.455675 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.455724 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.455785 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.455826 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-session\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.455865 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7745\" (UniqueName: \"kubernetes.io/projected/f71a6b20-21c9-4920-a510-a5817462836a-kube-api-access-v7745\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.455957 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.456012 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-template-error\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.456053 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.456104 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f71a6b20-21c9-4920-a510-a5817462836a-audit-dir\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.456146 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.456182 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-template-login\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.456252 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-audit-policies\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.458066 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.458299 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f71a6b20-21c9-4920-a510-a5817462836a-audit-dir\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.458388 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.458476 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-audit-policies\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.462637 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.463398 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.463833 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.463963 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.464257 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.464369 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.466405 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-template-login\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.466710 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-user-template-error\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.468046 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f71a6b20-21c9-4920-a510-a5817462836a-v4-0-config-system-session\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.487030 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7745\" (UniqueName: \"kubernetes.io/projected/f71a6b20-21c9-4920-a510-a5817462836a-kube-api-access-v7745\") pod \"oauth-openshift-6b788bb46c-9pxdc\" (UID: \"f71a6b20-21c9-4920-a510-a5817462836a\") " pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.520929 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.763750 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.789684 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:25:57 crc kubenswrapper[4966]: I1217 08:25:57.919686 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.080608 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.124904 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.365201 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.428457 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.461553 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.476342 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.482677 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.506535 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.551802 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.587148 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.699579 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.818061 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.839079 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="021d4c31-6de9-4e76-ac80-0f345b3d7f81" path="/var/lib/kubelet/pods/021d4c31-6de9-4e76-ac80-0f345b3d7f81/volumes" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.839843 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5c75559-5110-464f-893d-421f55a5318f" path="/var/lib/kubelet/pods/d5c75559-5110-464f-893d-421f55a5318f/volumes" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.889946 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.905770 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.918321 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.940650 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.954928 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6b788bb46c-9pxdc"] Dec 17 08:25:58 crc kubenswrapper[4966]: I1217 08:25:58.964826 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.096263 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.195054 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.273441 4966 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.285320 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.509843 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.559092 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.632775 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.642678 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6b788bb46c-9pxdc"] Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.666925 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.671418 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" event={"ID":"f71a6b20-21c9-4920-a510-a5817462836a","Type":"ContainerStarted","Data":"32602b5d7de2f1212b8e130c1fc5998b4b625c7f13752e7930f243868a7a9c45"} Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.712385 4966 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 17 08:25:59 crc kubenswrapper[4966]: I1217 08:25:59.906180 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.006504 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.136395 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.178383 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.180627 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.296564 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.448086 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.530116 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.597198 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.677029 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b788bb46c-9pxdc_f71a6b20-21c9-4920-a510-a5817462836a/oauth-openshift/0.log" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.677076 4966 generic.go:334] "Generic (PLEG): container finished" podID="f71a6b20-21c9-4920-a510-a5817462836a" containerID="530cb8332bf91ced558127a86fe9f3fe3c8604d4532845f3b3796afba55522ba" exitCode=255 Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.677099 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" event={"ID":"f71a6b20-21c9-4920-a510-a5817462836a","Type":"ContainerDied","Data":"530cb8332bf91ced558127a86fe9f3fe3c8604d4532845f3b3796afba55522ba"} Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.677494 4966 scope.go:117] "RemoveContainer" containerID="530cb8332bf91ced558127a86fe9f3fe3c8604d4532845f3b3796afba55522ba" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.696704 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.765825 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.791156 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.876289 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.917521 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 17 08:26:00 crc kubenswrapper[4966]: I1217 08:26:00.942960 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.187840 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.203845 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.284427 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.287150 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.295414 4966 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.360833 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.433240 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.440050 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.441490 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.452911 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.474143 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.482466 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.590520 4966 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.619722 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.685654 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b788bb46c-9pxdc_f71a6b20-21c9-4920-a510-a5817462836a/oauth-openshift/1.log" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.688955 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b788bb46c-9pxdc_f71a6b20-21c9-4920-a510-a5817462836a/oauth-openshift/0.log" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.689019 4966 generic.go:334] "Generic (PLEG): container finished" podID="f71a6b20-21c9-4920-a510-a5817462836a" containerID="da6159399a44b5150e8daad6bc8eb85ebe661511b3a41767aeffe500c288bac9" exitCode=255 Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.689057 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" event={"ID":"f71a6b20-21c9-4920-a510-a5817462836a","Type":"ContainerDied","Data":"da6159399a44b5150e8daad6bc8eb85ebe661511b3a41767aeffe500c288bac9"} Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.689106 4966 scope.go:117] "RemoveContainer" containerID="530cb8332bf91ced558127a86fe9f3fe3c8604d4532845f3b3796afba55522ba" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.690910 4966 scope.go:117] "RemoveContainer" containerID="da6159399a44b5150e8daad6bc8eb85ebe661511b3a41767aeffe500c288bac9" Dec 17 08:26:01 crc kubenswrapper[4966]: E1217 08:26:01.691218 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-6b788bb46c-9pxdc_openshift-authentication(f71a6b20-21c9-4920-a510-a5817462836a)\"" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" podUID="f71a6b20-21c9-4920-a510-a5817462836a" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.709033 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.722095 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.813464 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.847996 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.904423 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 17 08:26:01 crc kubenswrapper[4966]: I1217 08:26:01.938235 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.212600 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.218635 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.234662 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.287558 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.312285 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.378331 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.398948 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.452805 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.579162 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.609605 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.612864 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.634954 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.651559 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.653733 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.655046 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.696144 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b788bb46c-9pxdc_f71a6b20-21c9-4920-a510-a5817462836a/oauth-openshift/1.log" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.696642 4966 scope.go:117] "RemoveContainer" containerID="da6159399a44b5150e8daad6bc8eb85ebe661511b3a41767aeffe500c288bac9" Dec 17 08:26:02 crc kubenswrapper[4966]: E1217 08:26:02.696841 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-6b788bb46c-9pxdc_openshift-authentication(f71a6b20-21c9-4920-a510-a5817462836a)\"" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" podUID="f71a6b20-21c9-4920-a510-a5817462836a" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.732819 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 17 08:26:02 crc kubenswrapper[4966]: I1217 08:26:02.740195 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.043738 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.157094 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.236812 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.262269 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.273754 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.288816 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.296532 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.303939 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.326097 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.574490 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.737172 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.826359 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.868811 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.869085 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 17 08:26:03 crc kubenswrapper[4966]: I1217 08:26:03.964143 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.109586 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.110006 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.223408 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.441233 4966 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.450431 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.487313 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.560549 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.704589 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.740392 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.775710 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.811029 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 17 08:26:04 crc kubenswrapper[4966]: I1217 08:26:04.869759 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.092666 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.299832 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.330251 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.343786 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.354092 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.356169 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.415982 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.492815 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.545629 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.618441 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.679753 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.698257 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.790823 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.834151 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.939717 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.945753 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 17 08:26:05 crc kubenswrapper[4966]: I1217 08:26:05.991540 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.014744 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.073294 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.110077 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.137572 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.145910 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.169679 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.204106 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.268277 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.319838 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.320015 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.337472 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.344582 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.349033 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.357116 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.365011 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.468009 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.530093 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.553572 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.594165 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.678789 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.726509 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.760587 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.796699 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.834358 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.880988 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.901865 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.931516 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.933673 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 17 08:26:06 crc kubenswrapper[4966]: I1217 08:26:06.994280 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.007823 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.041672 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.067784 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.098022 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.176016 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.180030 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.250055 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.277014 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.277950 4966 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.278181 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://01f9f9b46c93b0fd3ed051c0452cdf4fdb69f5814410d5dddb6802b2ab199b90" gracePeriod=5 Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.311298 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.366686 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.382352 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.392361 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.545647 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.555051 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.579396 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.586655 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.707289 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.750765 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.764264 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.789982 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.790044 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.790682 4966 scope.go:117] "RemoveContainer" containerID="da6159399a44b5150e8daad6bc8eb85ebe661511b3a41767aeffe500c288bac9" Dec 17 08:26:07 crc kubenswrapper[4966]: E1217 08:26:07.791011 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-6b788bb46c-9pxdc_openshift-authentication(f71a6b20-21c9-4920-a510-a5817462836a)\"" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" podUID="f71a6b20-21c9-4920-a510-a5817462836a" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.832496 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.919178 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.950295 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 17 08:26:07 crc kubenswrapper[4966]: I1217 08:26:07.961328 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.123922 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.181705 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.290481 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.309985 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.455781 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.586475 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.779129 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.785981 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.850086 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 17 08:26:08 crc kubenswrapper[4966]: I1217 08:26:08.857624 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 17 08:26:09 crc kubenswrapper[4966]: I1217 08:26:09.085541 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 17 08:26:09 crc kubenswrapper[4966]: I1217 08:26:09.294086 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 17 08:26:09 crc kubenswrapper[4966]: I1217 08:26:09.597086 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 17 08:26:09 crc kubenswrapper[4966]: I1217 08:26:09.683912 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 17 08:26:09 crc kubenswrapper[4966]: I1217 08:26:09.702191 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 17 08:26:09 crc kubenswrapper[4966]: I1217 08:26:09.755798 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 17 08:26:09 crc kubenswrapper[4966]: I1217 08:26:09.763576 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 17 08:26:09 crc kubenswrapper[4966]: I1217 08:26:09.772811 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 17 08:26:09 crc kubenswrapper[4966]: I1217 08:26:09.986325 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 17 08:26:10 crc kubenswrapper[4966]: I1217 08:26:10.031563 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 17 08:26:10 crc kubenswrapper[4966]: I1217 08:26:10.077738 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 17 08:26:10 crc kubenswrapper[4966]: I1217 08:26:10.212449 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 17 08:26:10 crc kubenswrapper[4966]: I1217 08:26:10.337441 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 17 08:26:10 crc kubenswrapper[4966]: I1217 08:26:10.381060 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 17 08:26:10 crc kubenswrapper[4966]: I1217 08:26:10.400037 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 17 08:26:10 crc kubenswrapper[4966]: I1217 08:26:10.579278 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 17 08:26:10 crc kubenswrapper[4966]: I1217 08:26:10.989790 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 17 08:26:11 crc kubenswrapper[4966]: I1217 08:26:11.309804 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 17 08:26:11 crc kubenswrapper[4966]: I1217 08:26:11.429019 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 17 08:26:11 crc kubenswrapper[4966]: I1217 08:26:11.725812 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 17 08:26:11 crc kubenswrapper[4966]: I1217 08:26:11.955642 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 17 08:26:12 crc kubenswrapper[4966]: I1217 08:26:12.278794 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 17 08:26:12 crc kubenswrapper[4966]: I1217 08:26:12.751519 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 17 08:26:12 crc kubenswrapper[4966]: I1217 08:26:12.751633 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 17 08:26:12 crc kubenswrapper[4966]: I1217 08:26:12.758937 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 17 08:26:12 crc kubenswrapper[4966]: I1217 08:26:12.758986 4966 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="01f9f9b46c93b0fd3ed051c0452cdf4fdb69f5814410d5dddb6802b2ab199b90" exitCode=137 Dec 17 08:26:12 crc kubenswrapper[4966]: I1217 08:26:12.802855 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 17 08:26:12 crc kubenswrapper[4966]: I1217 08:26:12.864291 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 17 08:26:12 crc kubenswrapper[4966]: I1217 08:26:12.864373 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.059927 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.060152 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.060168 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.060191 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.060290 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.060082 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.060646 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.060747 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.060790 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.072003 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.161913 4966 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.161940 4966 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.161948 4966 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.161956 4966 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.161981 4966 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.766824 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.766939 4966 scope.go:117] "RemoveContainer" containerID="01f9f9b46c93b0fd3ed051c0452cdf4fdb69f5814410d5dddb6802b2ab199b90" Dec 17 08:26:13 crc kubenswrapper[4966]: I1217 08:26:13.766972 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 17 08:26:14 crc kubenswrapper[4966]: I1217 08:26:14.837044 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 17 08:26:20 crc kubenswrapper[4966]: I1217 08:26:20.830957 4966 scope.go:117] "RemoveContainer" containerID="da6159399a44b5150e8daad6bc8eb85ebe661511b3a41767aeffe500c288bac9" Dec 17 08:26:21 crc kubenswrapper[4966]: I1217 08:26:21.810117 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b788bb46c-9pxdc_f71a6b20-21c9-4920-a510-a5817462836a/oauth-openshift/1.log" Dec 17 08:26:21 crc kubenswrapper[4966]: I1217 08:26:21.810463 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" event={"ID":"f71a6b20-21c9-4920-a510-a5817462836a","Type":"ContainerStarted","Data":"9b044715551e7725a141b9faa7e28a1e4b17d377cd63127ae9b2a3348a6b72ba"} Dec 17 08:26:21 crc kubenswrapper[4966]: I1217 08:26:21.810939 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:26:21 crc kubenswrapper[4966]: I1217 08:26:21.817535 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" Dec 17 08:26:21 crc kubenswrapper[4966]: I1217 08:26:21.883597 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6b788bb46c-9pxdc" podStartSLOduration=63.883578522 podStartE2EDuration="1m3.883578522s" podCreationTimestamp="2025-12-17 08:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:26:21.850840165 +0000 UTC m=+317.395910117" watchObservedRunningTime="2025-12-17 08:26:21.883578522 +0000 UTC m=+317.428648474" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.045107 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hblng"] Dec 17 08:26:34 crc kubenswrapper[4966]: E1217 08:26:34.045853 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.045887 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.046019 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.046935 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.048967 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.058107 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hblng"] Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.246247 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b42edb-955a-4719-921d-2fc4528778e9-utilities\") pod \"certified-operators-hblng\" (UID: \"09b42edb-955a-4719-921d-2fc4528778e9\") " pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.246799 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b42edb-955a-4719-921d-2fc4528778e9-catalog-content\") pod \"certified-operators-hblng\" (UID: \"09b42edb-955a-4719-921d-2fc4528778e9\") " pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.247230 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xb7p\" (UniqueName: \"kubernetes.io/projected/09b42edb-955a-4719-921d-2fc4528778e9-kube-api-access-5xb7p\") pod \"certified-operators-hblng\" (UID: \"09b42edb-955a-4719-921d-2fc4528778e9\") " pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.348134 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b42edb-955a-4719-921d-2fc4528778e9-utilities\") pod \"certified-operators-hblng\" (UID: \"09b42edb-955a-4719-921d-2fc4528778e9\") " pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.348534 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b42edb-955a-4719-921d-2fc4528778e9-catalog-content\") pod \"certified-operators-hblng\" (UID: \"09b42edb-955a-4719-921d-2fc4528778e9\") " pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.348566 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b42edb-955a-4719-921d-2fc4528778e9-utilities\") pod \"certified-operators-hblng\" (UID: \"09b42edb-955a-4719-921d-2fc4528778e9\") " pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.348786 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xb7p\" (UniqueName: \"kubernetes.io/projected/09b42edb-955a-4719-921d-2fc4528778e9-kube-api-access-5xb7p\") pod \"certified-operators-hblng\" (UID: \"09b42edb-955a-4719-921d-2fc4528778e9\") " pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.349153 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b42edb-955a-4719-921d-2fc4528778e9-catalog-content\") pod \"certified-operators-hblng\" (UID: \"09b42edb-955a-4719-921d-2fc4528778e9\") " pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.368725 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xb7p\" (UniqueName: \"kubernetes.io/projected/09b42edb-955a-4719-921d-2fc4528778e9-kube-api-access-5xb7p\") pod \"certified-operators-hblng\" (UID: \"09b42edb-955a-4719-921d-2fc4528778e9\") " pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.663914 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:34 crc kubenswrapper[4966]: I1217 08:26:34.898118 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hblng"] Dec 17 08:26:34 crc kubenswrapper[4966]: W1217 08:26:34.909007 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09b42edb_955a_4719_921d_2fc4528778e9.slice/crio-52e66f9a4fe7a1625437a8e0c222affb82f96605552815f54efe1470a3c71075 WatchSource:0}: Error finding container 52e66f9a4fe7a1625437a8e0c222affb82f96605552815f54efe1470a3c71075: Status 404 returned error can't find the container with id 52e66f9a4fe7a1625437a8e0c222affb82f96605552815f54efe1470a3c71075 Dec 17 08:26:35 crc kubenswrapper[4966]: I1217 08:26:35.891179 4966 generic.go:334] "Generic (PLEG): container finished" podID="09b42edb-955a-4719-921d-2fc4528778e9" containerID="50a11789f116dce5c0aece477e837ae945b79305fd388bbfabd9fac0fafa5ac3" exitCode=0 Dec 17 08:26:35 crc kubenswrapper[4966]: I1217 08:26:35.891262 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hblng" event={"ID":"09b42edb-955a-4719-921d-2fc4528778e9","Type":"ContainerDied","Data":"50a11789f116dce5c0aece477e837ae945b79305fd388bbfabd9fac0fafa5ac3"} Dec 17 08:26:35 crc kubenswrapper[4966]: I1217 08:26:35.891553 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hblng" event={"ID":"09b42edb-955a-4719-921d-2fc4528778e9","Type":"ContainerStarted","Data":"52e66f9a4fe7a1625437a8e0c222affb82f96605552815f54efe1470a3c71075"} Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.644155 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gvr6s"] Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.645632 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.648597 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.659006 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gvr6s"] Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.783471 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7chsc\" (UniqueName: \"kubernetes.io/projected/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-kube-api-access-7chsc\") pod \"redhat-operators-gvr6s\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.783522 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-utilities\") pod \"redhat-operators-gvr6s\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.783543 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-catalog-content\") pod \"redhat-operators-gvr6s\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.886923 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7chsc\" (UniqueName: \"kubernetes.io/projected/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-kube-api-access-7chsc\") pod \"redhat-operators-gvr6s\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.887009 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-utilities\") pod \"redhat-operators-gvr6s\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.887038 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-catalog-content\") pod \"redhat-operators-gvr6s\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.887900 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-utilities\") pod \"redhat-operators-gvr6s\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.888173 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-catalog-content\") pod \"redhat-operators-gvr6s\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.897812 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hblng" event={"ID":"09b42edb-955a-4719-921d-2fc4528778e9","Type":"ContainerStarted","Data":"5fd8093a32f206c95cee7e8edc8c3610c66a88d174ec6f307966c6d0a511c0b3"} Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.925407 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7chsc\" (UniqueName: \"kubernetes.io/projected/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-kube-api-access-7chsc\") pod \"redhat-operators-gvr6s\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:36 crc kubenswrapper[4966]: I1217 08:26:36.962094 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:37 crc kubenswrapper[4966]: I1217 08:26:37.183361 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gvr6s"] Dec 17 08:26:37 crc kubenswrapper[4966]: W1217 08:26:37.214568 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9db233fb_f3e8_4b7d_a1ee_2f0fa974f81c.slice/crio-2fd4c5e1c7953e3744400d7c08094799f290ce4ed569f7ee8c672d09febcfcdb WatchSource:0}: Error finding container 2fd4c5e1c7953e3744400d7c08094799f290ce4ed569f7ee8c672d09febcfcdb: Status 404 returned error can't find the container with id 2fd4c5e1c7953e3744400d7c08094799f290ce4ed569f7ee8c672d09febcfcdb Dec 17 08:26:37 crc kubenswrapper[4966]: I1217 08:26:37.905449 4966 generic.go:334] "Generic (PLEG): container finished" podID="09b42edb-955a-4719-921d-2fc4528778e9" containerID="5fd8093a32f206c95cee7e8edc8c3610c66a88d174ec6f307966c6d0a511c0b3" exitCode=0 Dec 17 08:26:37 crc kubenswrapper[4966]: I1217 08:26:37.905519 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hblng" event={"ID":"09b42edb-955a-4719-921d-2fc4528778e9","Type":"ContainerDied","Data":"5fd8093a32f206c95cee7e8edc8c3610c66a88d174ec6f307966c6d0a511c0b3"} Dec 17 08:26:37 crc kubenswrapper[4966]: I1217 08:26:37.907667 4966 generic.go:334] "Generic (PLEG): container finished" podID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerID="9fa20c5d41bb583227a1ef0070db1556f83021a47856052d810bb227051d36f8" exitCode=0 Dec 17 08:26:37 crc kubenswrapper[4966]: I1217 08:26:37.907718 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvr6s" event={"ID":"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c","Type":"ContainerDied","Data":"9fa20c5d41bb583227a1ef0070db1556f83021a47856052d810bb227051d36f8"} Dec 17 08:26:37 crc kubenswrapper[4966]: I1217 08:26:37.907743 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvr6s" event={"ID":"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c","Type":"ContainerStarted","Data":"2fd4c5e1c7953e3744400d7c08094799f290ce4ed569f7ee8c672d09febcfcdb"} Dec 17 08:26:38 crc kubenswrapper[4966]: I1217 08:26:38.916764 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hblng" event={"ID":"09b42edb-955a-4719-921d-2fc4528778e9","Type":"ContainerStarted","Data":"985cb2c4afe41aff156448928e37a3d7eda77e65b9f697c1d689cb761275e880"} Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.308122 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hblng" podStartSLOduration=2.716627866 podStartE2EDuration="5.308106244s" podCreationTimestamp="2025-12-17 08:26:34 +0000 UTC" firstStartedPulling="2025-12-17 08:26:35.892860602 +0000 UTC m=+331.437930544" lastFinishedPulling="2025-12-17 08:26:38.48433898 +0000 UTC m=+334.029408922" observedRunningTime="2025-12-17 08:26:38.937429386 +0000 UTC m=+334.482499328" watchObservedRunningTime="2025-12-17 08:26:39.308106244 +0000 UTC m=+334.853176186" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.308844 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v6b7k"] Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.309107 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" podUID="41dbc543-cace-4c8f-8db3-e07b984c0919" containerName="controller-manager" containerID="cri-o://dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec" gracePeriod=30 Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.451401 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt"] Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.451642 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" podUID="1b60ed2c-77ac-4cbc-9b57-48ed79419073" containerName="route-controller-manager" containerID="cri-o://4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527" gracePeriod=30 Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.728196 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.793762 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.828859 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-config\") pod \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.828939 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41dbc543-cace-4c8f-8db3-e07b984c0919-serving-cert\") pod \"41dbc543-cace-4c8f-8db3-e07b984c0919\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.828982 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-config\") pod \"41dbc543-cace-4c8f-8db3-e07b984c0919\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.829024 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-client-ca\") pod \"41dbc543-cace-4c8f-8db3-e07b984c0919\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.829069 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-545cz\" (UniqueName: \"kubernetes.io/projected/1b60ed2c-77ac-4cbc-9b57-48ed79419073-kube-api-access-545cz\") pod \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.829093 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-client-ca\") pod \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.829116 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-proxy-ca-bundles\") pod \"41dbc543-cace-4c8f-8db3-e07b984c0919\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.829143 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbzqw\" (UniqueName: \"kubernetes.io/projected/41dbc543-cace-4c8f-8db3-e07b984c0919-kube-api-access-sbzqw\") pod \"41dbc543-cace-4c8f-8db3-e07b984c0919\" (UID: \"41dbc543-cace-4c8f-8db3-e07b984c0919\") " Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.829166 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b60ed2c-77ac-4cbc-9b57-48ed79419073-serving-cert\") pod \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\" (UID: \"1b60ed2c-77ac-4cbc-9b57-48ed79419073\") " Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.830692 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-config" (OuterVolumeSpecName: "config") pod "1b60ed2c-77ac-4cbc-9b57-48ed79419073" (UID: "1b60ed2c-77ac-4cbc-9b57-48ed79419073"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.830745 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-client-ca" (OuterVolumeSpecName: "client-ca") pod "41dbc543-cace-4c8f-8db3-e07b984c0919" (UID: "41dbc543-cace-4c8f-8db3-e07b984c0919"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.830767 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-config" (OuterVolumeSpecName: "config") pod "41dbc543-cace-4c8f-8db3-e07b984c0919" (UID: "41dbc543-cace-4c8f-8db3-e07b984c0919"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.831186 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-client-ca" (OuterVolumeSpecName: "client-ca") pod "1b60ed2c-77ac-4cbc-9b57-48ed79419073" (UID: "1b60ed2c-77ac-4cbc-9b57-48ed79419073"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.831698 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "41dbc543-cace-4c8f-8db3-e07b984c0919" (UID: "41dbc543-cace-4c8f-8db3-e07b984c0919"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.835243 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41dbc543-cace-4c8f-8db3-e07b984c0919-kube-api-access-sbzqw" (OuterVolumeSpecName: "kube-api-access-sbzqw") pod "41dbc543-cace-4c8f-8db3-e07b984c0919" (UID: "41dbc543-cace-4c8f-8db3-e07b984c0919"). InnerVolumeSpecName "kube-api-access-sbzqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.835662 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b60ed2c-77ac-4cbc-9b57-48ed79419073-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1b60ed2c-77ac-4cbc-9b57-48ed79419073" (UID: "1b60ed2c-77ac-4cbc-9b57-48ed79419073"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.840321 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dbc543-cace-4c8f-8db3-e07b984c0919-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "41dbc543-cace-4c8f-8db3-e07b984c0919" (UID: "41dbc543-cace-4c8f-8db3-e07b984c0919"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.843075 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b60ed2c-77ac-4cbc-9b57-48ed79419073-kube-api-access-545cz" (OuterVolumeSpecName: "kube-api-access-545cz") pod "1b60ed2c-77ac-4cbc-9b57-48ed79419073" (UID: "1b60ed2c-77ac-4cbc-9b57-48ed79419073"). InnerVolumeSpecName "kube-api-access-545cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.922562 4966 generic.go:334] "Generic (PLEG): container finished" podID="1b60ed2c-77ac-4cbc-9b57-48ed79419073" containerID="4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527" exitCode=0 Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.922600 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" event={"ID":"1b60ed2c-77ac-4cbc-9b57-48ed79419073","Type":"ContainerDied","Data":"4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527"} Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.922640 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.922660 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt" event={"ID":"1b60ed2c-77ac-4cbc-9b57-48ed79419073","Type":"ContainerDied","Data":"41acd75f509ba5f4494933aef48f5fab95fc579e70b34fd723cc9abcbdb0333c"} Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.922682 4966 scope.go:117] "RemoveContainer" containerID="4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.924307 4966 generic.go:334] "Generic (PLEG): container finished" podID="41dbc543-cace-4c8f-8db3-e07b984c0919" containerID="dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec" exitCode=0 Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.924383 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.924422 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" event={"ID":"41dbc543-cace-4c8f-8db3-e07b984c0919","Type":"ContainerDied","Data":"dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec"} Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.924454 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v6b7k" event={"ID":"41dbc543-cace-4c8f-8db3-e07b984c0919","Type":"ContainerDied","Data":"71a67e5f2925413a5df9fe692efd5a759414ef5c38425da0513628cddbca3f63"} Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.928271 4966 generic.go:334] "Generic (PLEG): container finished" podID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerID="0277cb0975789b6b645e1f55ebc25cf546ac333381c0d31cc605fb271772ece3" exitCode=0 Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.934098 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvr6s" event={"ID":"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c","Type":"ContainerDied","Data":"0277cb0975789b6b645e1f55ebc25cf546ac333381c0d31cc605fb271772ece3"} Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.934771 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41dbc543-cace-4c8f-8db3-e07b984c0919-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.934793 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.934804 4966 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-client-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.934816 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-545cz\" (UniqueName: \"kubernetes.io/projected/1b60ed2c-77ac-4cbc-9b57-48ed79419073-kube-api-access-545cz\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.934828 4966 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-client-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.934839 4966 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41dbc543-cace-4c8f-8db3-e07b984c0919-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.934851 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbzqw\" (UniqueName: \"kubernetes.io/projected/41dbc543-cace-4c8f-8db3-e07b984c0919-kube-api-access-sbzqw\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.934862 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b60ed2c-77ac-4cbc-9b57-48ed79419073-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.935577 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b60ed2c-77ac-4cbc-9b57-48ed79419073-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.954906 4966 scope.go:117] "RemoveContainer" containerID="4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527" Dec 17 08:26:39 crc kubenswrapper[4966]: E1217 08:26:39.958131 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527\": container with ID starting with 4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527 not found: ID does not exist" containerID="4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.958178 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527"} err="failed to get container status \"4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527\": rpc error: code = NotFound desc = could not find container \"4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527\": container with ID starting with 4e79fd180471b6fc9c00e5011cf6f1d3d425ee2aa7b1a548194f5aa5e9c8e527 not found: ID does not exist" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.958203 4966 scope.go:117] "RemoveContainer" containerID="dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.973904 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt"] Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.979982 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gzzdt"] Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.987120 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v6b7k"] Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.991039 4966 scope.go:117] "RemoveContainer" containerID="dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec" Dec 17 08:26:39 crc kubenswrapper[4966]: E1217 08:26:39.991974 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec\": container with ID starting with dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec not found: ID does not exist" containerID="dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.992013 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec"} err="failed to get container status \"dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec\": rpc error: code = NotFound desc = could not find container \"dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec\": container with ID starting with dca3715843e1fff8fc3887fd289fb9084cb0d65666db8e15c840b95ce17714ec not found: ID does not exist" Dec 17 08:26:39 crc kubenswrapper[4966]: I1217 08:26:39.993944 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v6b7k"] Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.516355 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc"] Dec 17 08:26:40 crc kubenswrapper[4966]: E1217 08:26:40.516585 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dbc543-cace-4c8f-8db3-e07b984c0919" containerName="controller-manager" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.516601 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dbc543-cace-4c8f-8db3-e07b984c0919" containerName="controller-manager" Dec 17 08:26:40 crc kubenswrapper[4966]: E1217 08:26:40.516628 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b60ed2c-77ac-4cbc-9b57-48ed79419073" containerName="route-controller-manager" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.516636 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b60ed2c-77ac-4cbc-9b57-48ed79419073" containerName="route-controller-manager" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.516738 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b60ed2c-77ac-4cbc-9b57-48ed79419073" containerName="route-controller-manager" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.516761 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dbc543-cace-4c8f-8db3-e07b984c0919" containerName="controller-manager" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.517231 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.521608 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.521807 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.521953 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.522057 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.527190 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc"] Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.527956 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.527959 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.541982 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82143ef0-c796-42df-84c9-5814f7e77d9f-config\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.542325 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncdsw\" (UniqueName: \"kubernetes.io/projected/82143ef0-c796-42df-84c9-5814f7e77d9f-kube-api-access-ncdsw\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.542446 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82143ef0-c796-42df-84c9-5814f7e77d9f-client-ca\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.542592 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82143ef0-c796-42df-84c9-5814f7e77d9f-serving-cert\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.644350 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82143ef0-c796-42df-84c9-5814f7e77d9f-config\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.644451 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82143ef0-c796-42df-84c9-5814f7e77d9f-client-ca\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.644475 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncdsw\" (UniqueName: \"kubernetes.io/projected/82143ef0-c796-42df-84c9-5814f7e77d9f-kube-api-access-ncdsw\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.644516 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82143ef0-c796-42df-84c9-5814f7e77d9f-serving-cert\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.645590 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/82143ef0-c796-42df-84c9-5814f7e77d9f-client-ca\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.645650 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82143ef0-c796-42df-84c9-5814f7e77d9f-config\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.648318 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82143ef0-c796-42df-84c9-5814f7e77d9f-serving-cert\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.663136 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncdsw\" (UniqueName: \"kubernetes.io/projected/82143ef0-c796-42df-84c9-5814f7e77d9f-kube-api-access-ncdsw\") pod \"route-controller-manager-5876db6c88-chczc\" (UID: \"82143ef0-c796-42df-84c9-5814f7e77d9f\") " pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.836557 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b60ed2c-77ac-4cbc-9b57-48ed79419073" path="/var/lib/kubelet/pods/1b60ed2c-77ac-4cbc-9b57-48ed79419073/volumes" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.837344 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41dbc543-cace-4c8f-8db3-e07b984c0919" path="/var/lib/kubelet/pods/41dbc543-cace-4c8f-8db3-e07b984c0919/volumes" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.841375 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.948675 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvr6s" event={"ID":"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c","Type":"ContainerStarted","Data":"8a03fe11e7a3d7a669a8a140e8147a154fad325a66c3d030d96299652f39c054"} Dec 17 08:26:40 crc kubenswrapper[4966]: I1217 08:26:40.980035 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gvr6s" podStartSLOduration=2.264068234 podStartE2EDuration="4.9800147s" podCreationTimestamp="2025-12-17 08:26:36 +0000 UTC" firstStartedPulling="2025-12-17 08:26:37.908918796 +0000 UTC m=+333.453988748" lastFinishedPulling="2025-12-17 08:26:40.624865272 +0000 UTC m=+336.169935214" observedRunningTime="2025-12-17 08:26:40.97858051 +0000 UTC m=+336.523650462" watchObservedRunningTime="2025-12-17 08:26:40.9800147 +0000 UTC m=+336.525084662" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.321112 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc"] Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.515565 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-64585bb48f-stzs9"] Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.516848 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.519622 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.519753 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.519987 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.521402 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.525711 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64585bb48f-stzs9"] Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.529470 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.530305 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.532827 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.554663 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bw2k\" (UniqueName: \"kubernetes.io/projected/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-kube-api-access-6bw2k\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.554779 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-client-ca\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.554811 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-config\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.554849 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-proxy-ca-bundles\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.554896 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-serving-cert\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.655634 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bw2k\" (UniqueName: \"kubernetes.io/projected/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-kube-api-access-6bw2k\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.655729 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-client-ca\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.655759 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-config\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.655786 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-proxy-ca-bundles\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.655808 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-serving-cert\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.657258 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-client-ca\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.658101 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-proxy-ca-bundles\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.658159 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-config\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.666857 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-serving-cert\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.680716 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bw2k\" (UniqueName: \"kubernetes.io/projected/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-kube-api-access-6bw2k\") pod \"controller-manager-64585bb48f-stzs9\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.837679 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.862422 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7fjhc"] Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.863616 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.872079 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7fjhc"] Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.958553 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq94s\" (UniqueName: \"kubernetes.io/projected/69c77d23-84b6-4893-a6c6-e8908843c1cf-kube-api-access-fq94s\") pod \"certified-operators-7fjhc\" (UID: \"69c77d23-84b6-4893-a6c6-e8908843c1cf\") " pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.958806 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69c77d23-84b6-4893-a6c6-e8908843c1cf-catalog-content\") pod \"certified-operators-7fjhc\" (UID: \"69c77d23-84b6-4893-a6c6-e8908843c1cf\") " pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.958825 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69c77d23-84b6-4893-a6c6-e8908843c1cf-utilities\") pod \"certified-operators-7fjhc\" (UID: \"69c77d23-84b6-4893-a6c6-e8908843c1cf\") " pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.964859 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" event={"ID":"82143ef0-c796-42df-84c9-5814f7e77d9f","Type":"ContainerStarted","Data":"3d93c48e8c525b857b098004be331acac46750df4cb3fb61feafda578d0d9b10"} Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.964910 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" event={"ID":"82143ef0-c796-42df-84c9-5814f7e77d9f","Type":"ContainerStarted","Data":"639313db7bbba9224870d7865e4cf8253ff23709c5329c5fd6d5ef85f6bff3c4"} Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.965227 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:41 crc kubenswrapper[4966]: I1217 08:26:41.990467 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" podStartSLOduration=2.9904463310000002 podStartE2EDuration="2.990446331s" podCreationTimestamp="2025-12-17 08:26:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:26:41.982824145 +0000 UTC m=+337.527894087" watchObservedRunningTime="2025-12-17 08:26:41.990446331 +0000 UTC m=+337.535516283" Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.047406 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5876db6c88-chczc" Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.060482 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69c77d23-84b6-4893-a6c6-e8908843c1cf-catalog-content\") pod \"certified-operators-7fjhc\" (UID: \"69c77d23-84b6-4893-a6c6-e8908843c1cf\") " pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.060523 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69c77d23-84b6-4893-a6c6-e8908843c1cf-utilities\") pod \"certified-operators-7fjhc\" (UID: \"69c77d23-84b6-4893-a6c6-e8908843c1cf\") " pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.060630 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq94s\" (UniqueName: \"kubernetes.io/projected/69c77d23-84b6-4893-a6c6-e8908843c1cf-kube-api-access-fq94s\") pod \"certified-operators-7fjhc\" (UID: \"69c77d23-84b6-4893-a6c6-e8908843c1cf\") " pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.062322 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69c77d23-84b6-4893-a6c6-e8908843c1cf-catalog-content\") pod \"certified-operators-7fjhc\" (UID: \"69c77d23-84b6-4893-a6c6-e8908843c1cf\") " pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.062604 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69c77d23-84b6-4893-a6c6-e8908843c1cf-utilities\") pod \"certified-operators-7fjhc\" (UID: \"69c77d23-84b6-4893-a6c6-e8908843c1cf\") " pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.063552 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64585bb48f-stzs9"] Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.083153 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq94s\" (UniqueName: \"kubernetes.io/projected/69c77d23-84b6-4893-a6c6-e8908843c1cf-kube-api-access-fq94s\") pod \"certified-operators-7fjhc\" (UID: \"69c77d23-84b6-4893-a6c6-e8908843c1cf\") " pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.181240 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.396067 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7fjhc"] Dec 17 08:26:42 crc kubenswrapper[4966]: W1217 08:26:42.402949 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69c77d23_84b6_4893_a6c6_e8908843c1cf.slice/crio-99966fd92505cf276b659b38dabd17af2d9c2265e6e07ad725d1e647ffd77f24 WatchSource:0}: Error finding container 99966fd92505cf276b659b38dabd17af2d9c2265e6e07ad725d1e647ffd77f24: Status 404 returned error can't find the container with id 99966fd92505cf276b659b38dabd17af2d9c2265e6e07ad725d1e647ffd77f24 Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.968803 4966 generic.go:334] "Generic (PLEG): container finished" podID="69c77d23-84b6-4893-a6c6-e8908843c1cf" containerID="6e1c3fb38cdb93b5f43823ce417bf76584c5485ca4d9db1a5d6d3b12ab0fb354" exitCode=0 Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.968912 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fjhc" event={"ID":"69c77d23-84b6-4893-a6c6-e8908843c1cf","Type":"ContainerDied","Data":"6e1c3fb38cdb93b5f43823ce417bf76584c5485ca4d9db1a5d6d3b12ab0fb354"} Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.968972 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fjhc" event={"ID":"69c77d23-84b6-4893-a6c6-e8908843c1cf","Type":"ContainerStarted","Data":"99966fd92505cf276b659b38dabd17af2d9c2265e6e07ad725d1e647ffd77f24"} Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.970637 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" event={"ID":"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2","Type":"ContainerStarted","Data":"1a0c4b4f11c39c6d66941a957903608f3dbc7f965cc926cc62e99e68e571a31f"} Dec 17 08:26:42 crc kubenswrapper[4966]: I1217 08:26:42.970667 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" event={"ID":"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2","Type":"ContainerStarted","Data":"9536e7cb7e84fb58ecf8cbb50cc1ace9b57f218932fdfa39e1eb254e28e457b3"} Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.442485 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" podStartSLOduration=4.442466002 podStartE2EDuration="4.442466002s" podCreationTimestamp="2025-12-17 08:26:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:26:43.032033131 +0000 UTC m=+338.577103083" watchObservedRunningTime="2025-12-17 08:26:43.442466002 +0000 UTC m=+338.987535944" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.443429 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pv2kq"] Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.444334 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.461688 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pv2kq"] Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.505491 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-utilities\") pod \"redhat-operators-pv2kq\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.505577 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scnwr\" (UniqueName: \"kubernetes.io/projected/982c7add-45d7-451e-8e31-3050c85eaa89-kube-api-access-scnwr\") pod \"redhat-operators-pv2kq\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.505682 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-catalog-content\") pod \"redhat-operators-pv2kq\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.607062 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-utilities\") pod \"redhat-operators-pv2kq\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.607143 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scnwr\" (UniqueName: \"kubernetes.io/projected/982c7add-45d7-451e-8e31-3050c85eaa89-kube-api-access-scnwr\") pod \"redhat-operators-pv2kq\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.607202 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-catalog-content\") pod \"redhat-operators-pv2kq\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.607853 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-utilities\") pod \"redhat-operators-pv2kq\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.607971 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-catalog-content\") pod \"redhat-operators-pv2kq\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.636814 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scnwr\" (UniqueName: \"kubernetes.io/projected/982c7add-45d7-451e-8e31-3050c85eaa89-kube-api-access-scnwr\") pod \"redhat-operators-pv2kq\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.758292 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.982205 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fjhc" event={"ID":"69c77d23-84b6-4893-a6c6-e8908843c1cf","Type":"ContainerStarted","Data":"dea75904179a33fd0a91eae29953bd75cb11374c6d3e9fef39ea6222cf74ab42"} Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.983083 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:43 crc kubenswrapper[4966]: I1217 08:26:43.993127 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:26:44 crc kubenswrapper[4966]: I1217 08:26:44.187423 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pv2kq"] Dec 17 08:26:44 crc kubenswrapper[4966]: I1217 08:26:44.665189 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:44 crc kubenswrapper[4966]: I1217 08:26:44.665604 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:44 crc kubenswrapper[4966]: I1217 08:26:44.721088 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:44 crc kubenswrapper[4966]: I1217 08:26:44.987687 4966 generic.go:334] "Generic (PLEG): container finished" podID="69c77d23-84b6-4893-a6c6-e8908843c1cf" containerID="dea75904179a33fd0a91eae29953bd75cb11374c6d3e9fef39ea6222cf74ab42" exitCode=0 Dec 17 08:26:44 crc kubenswrapper[4966]: I1217 08:26:44.987751 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fjhc" event={"ID":"69c77d23-84b6-4893-a6c6-e8908843c1cf","Type":"ContainerDied","Data":"dea75904179a33fd0a91eae29953bd75cb11374c6d3e9fef39ea6222cf74ab42"} Dec 17 08:26:44 crc kubenswrapper[4966]: I1217 08:26:44.992263 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pv2kq" event={"ID":"982c7add-45d7-451e-8e31-3050c85eaa89","Type":"ContainerStarted","Data":"c683ad190b7b29e464f98c00b8cd83423cf60ace24621230effac6569365e1bf"} Dec 17 08:26:44 crc kubenswrapper[4966]: I1217 08:26:44.992296 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pv2kq" event={"ID":"982c7add-45d7-451e-8e31-3050c85eaa89","Type":"ContainerStarted","Data":"ca7573a7d922fcf3440d801c4e853e1e8024189d0ea972a8773981b0532e1ee3"} Dec 17 08:26:45 crc kubenswrapper[4966]: I1217 08:26:45.056988 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hblng" Dec 17 08:26:45 crc kubenswrapper[4966]: I1217 08:26:45.844194 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nn68p"] Dec 17 08:26:45 crc kubenswrapper[4966]: I1217 08:26:45.845326 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:45 crc kubenswrapper[4966]: I1217 08:26:45.860126 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nn68p"] Dec 17 08:26:45 crc kubenswrapper[4966]: I1217 08:26:45.938129 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd-catalog-content\") pod \"certified-operators-nn68p\" (UID: \"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd\") " pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:45 crc kubenswrapper[4966]: I1217 08:26:45.938186 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz89n\" (UniqueName: \"kubernetes.io/projected/4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd-kube-api-access-zz89n\") pod \"certified-operators-nn68p\" (UID: \"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd\") " pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:45 crc kubenswrapper[4966]: I1217 08:26:45.938284 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd-utilities\") pod \"certified-operators-nn68p\" (UID: \"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd\") " pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:45 crc kubenswrapper[4966]: I1217 08:26:45.996844 4966 generic.go:334] "Generic (PLEG): container finished" podID="982c7add-45d7-451e-8e31-3050c85eaa89" containerID="c683ad190b7b29e464f98c00b8cd83423cf60ace24621230effac6569365e1bf" exitCode=0 Dec 17 08:26:45 crc kubenswrapper[4966]: I1217 08:26:45.998079 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pv2kq" event={"ID":"982c7add-45d7-451e-8e31-3050c85eaa89","Type":"ContainerDied","Data":"c683ad190b7b29e464f98c00b8cd83423cf60ace24621230effac6569365e1bf"} Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.040095 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd-utilities\") pod \"certified-operators-nn68p\" (UID: \"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd\") " pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.040252 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd-catalog-content\") pod \"certified-operators-nn68p\" (UID: \"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd\") " pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.040351 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz89n\" (UniqueName: \"kubernetes.io/projected/4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd-kube-api-access-zz89n\") pod \"certified-operators-nn68p\" (UID: \"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd\") " pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.040966 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd-catalog-content\") pod \"certified-operators-nn68p\" (UID: \"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd\") " pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.041097 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd-utilities\") pod \"certified-operators-nn68p\" (UID: \"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd\") " pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.074049 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz89n\" (UniqueName: \"kubernetes.io/projected/4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd-kube-api-access-zz89n\") pod \"certified-operators-nn68p\" (UID: \"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd\") " pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.161097 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.441270 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-db25g"] Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.442472 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.456784 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-db25g"] Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.549669 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-utilities\") pod \"redhat-operators-db25g\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.550035 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-catalog-content\") pod \"redhat-operators-db25g\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.550197 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzcxz\" (UniqueName: \"kubernetes.io/projected/2cd4b4c7-507c-4f2c-a657-b579d039441b-kube-api-access-pzcxz\") pod \"redhat-operators-db25g\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.615032 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nn68p"] Dec 17 08:26:46 crc kubenswrapper[4966]: W1217 08:26:46.621640 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d6d02de_ba82_48f6_aaaf_01e6a65b0dcd.slice/crio-c878dbadc6228ca0de4cf46d851df4e3b8cfb5e956e748426983f3b58384bbdc WatchSource:0}: Error finding container c878dbadc6228ca0de4cf46d851df4e3b8cfb5e956e748426983f3b58384bbdc: Status 404 returned error can't find the container with id c878dbadc6228ca0de4cf46d851df4e3b8cfb5e956e748426983f3b58384bbdc Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.650966 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-catalog-content\") pod \"redhat-operators-db25g\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.651169 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzcxz\" (UniqueName: \"kubernetes.io/projected/2cd4b4c7-507c-4f2c-a657-b579d039441b-kube-api-access-pzcxz\") pod \"redhat-operators-db25g\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.651318 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-utilities\") pod \"redhat-operators-db25g\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.651989 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-utilities\") pod \"redhat-operators-db25g\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.652439 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-catalog-content\") pod \"redhat-operators-db25g\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.670917 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzcxz\" (UniqueName: \"kubernetes.io/projected/2cd4b4c7-507c-4f2c-a657-b579d039441b-kube-api-access-pzcxz\") pod \"redhat-operators-db25g\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.768272 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.963969 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:46 crc kubenswrapper[4966]: I1217 08:26:46.964367 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:47 crc kubenswrapper[4966]: I1217 08:26:47.010083 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:47 crc kubenswrapper[4966]: I1217 08:26:47.011407 4966 generic.go:334] "Generic (PLEG): container finished" podID="4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd" containerID="6a154b6794b10ced9d2185ff377663579ec8a78c28d6be6979d8e17f4f75cbbb" exitCode=0 Dec 17 08:26:47 crc kubenswrapper[4966]: I1217 08:26:47.011468 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn68p" event={"ID":"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd","Type":"ContainerDied","Data":"6a154b6794b10ced9d2185ff377663579ec8a78c28d6be6979d8e17f4f75cbbb"} Dec 17 08:26:47 crc kubenswrapper[4966]: I1217 08:26:47.011538 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn68p" event={"ID":"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd","Type":"ContainerStarted","Data":"c878dbadc6228ca0de4cf46d851df4e3b8cfb5e956e748426983f3b58384bbdc"} Dec 17 08:26:47 crc kubenswrapper[4966]: I1217 08:26:47.015717 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fjhc" event={"ID":"69c77d23-84b6-4893-a6c6-e8908843c1cf","Type":"ContainerStarted","Data":"5c0be004b9cb8a75315bc21a4adcbf425d20c7cf2302623b3f9bb6a68a813577"} Dec 17 08:26:47 crc kubenswrapper[4966]: I1217 08:26:47.063216 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7fjhc" podStartSLOduration=2.748960646 podStartE2EDuration="6.063197833s" podCreationTimestamp="2025-12-17 08:26:41 +0000 UTC" firstStartedPulling="2025-12-17 08:26:42.969839583 +0000 UTC m=+338.514909525" lastFinishedPulling="2025-12-17 08:26:46.28407677 +0000 UTC m=+341.829146712" observedRunningTime="2025-12-17 08:26:47.060734243 +0000 UTC m=+342.605804185" watchObservedRunningTime="2025-12-17 08:26:47.063197833 +0000 UTC m=+342.608267775" Dec 17 08:26:47 crc kubenswrapper[4966]: I1217 08:26:47.071662 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 08:26:47 crc kubenswrapper[4966]: I1217 08:26:47.231060 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-db25g"] Dec 17 08:26:48 crc kubenswrapper[4966]: I1217 08:26:48.019536 4966 generic.go:334] "Generic (PLEG): container finished" podID="982c7add-45d7-451e-8e31-3050c85eaa89" containerID="9e94849486ea9a5989a2791c83a8f4292e223745832f1d41d1f51fe1da047e51" exitCode=0 Dec 17 08:26:48 crc kubenswrapper[4966]: I1217 08:26:48.019589 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pv2kq" event={"ID":"982c7add-45d7-451e-8e31-3050c85eaa89","Type":"ContainerDied","Data":"9e94849486ea9a5989a2791c83a8f4292e223745832f1d41d1f51fe1da047e51"} Dec 17 08:26:48 crc kubenswrapper[4966]: I1217 08:26:48.032577 4966 generic.go:334] "Generic (PLEG): container finished" podID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerID="50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca" exitCode=0 Dec 17 08:26:48 crc kubenswrapper[4966]: I1217 08:26:48.032686 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db25g" event={"ID":"2cd4b4c7-507c-4f2c-a657-b579d039441b","Type":"ContainerDied","Data":"50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca"} Dec 17 08:26:48 crc kubenswrapper[4966]: I1217 08:26:48.032707 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db25g" event={"ID":"2cd4b4c7-507c-4f2c-a657-b579d039441b","Type":"ContainerStarted","Data":"b4c9c3a10f088d00e219badf94a114b66b63bdad7cc3df0ddc2ddfe3137b88aa"} Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.040078 4966 generic.go:334] "Generic (PLEG): container finished" podID="4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd" containerID="5d68ad14fe39b54d19755564bbd9b74c96d3e39c7b8e297dd6a131d1bcd3fc3d" exitCode=0 Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.040124 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn68p" event={"ID":"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd","Type":"ContainerDied","Data":"5d68ad14fe39b54d19755564bbd9b74c96d3e39c7b8e297dd6a131d1bcd3fc3d"} Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.045001 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pv2kq" event={"ID":"982c7add-45d7-451e-8e31-3050c85eaa89","Type":"ContainerStarted","Data":"1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790"} Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.080411 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pv2kq" podStartSLOduration=3.647235665 podStartE2EDuration="6.080383158s" podCreationTimestamp="2025-12-17 08:26:43 +0000 UTC" firstStartedPulling="2025-12-17 08:26:45.998862609 +0000 UTC m=+341.543932551" lastFinishedPulling="2025-12-17 08:26:48.432010092 +0000 UTC m=+343.977080044" observedRunningTime="2025-12-17 08:26:49.079494333 +0000 UTC m=+344.624564285" watchObservedRunningTime="2025-12-17 08:26:49.080383158 +0000 UTC m=+344.625453130" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.573981 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xnlzx"] Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.575000 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.577391 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xnlzx"] Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.674657 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9szt\" (UniqueName: \"kubernetes.io/projected/a0309136-eab0-4861-8b0f-b0d9e783bb4c-kube-api-access-r9szt\") pod \"certified-operators-xnlzx\" (UID: \"a0309136-eab0-4861-8b0f-b0d9e783bb4c\") " pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.674739 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0309136-eab0-4861-8b0f-b0d9e783bb4c-catalog-content\") pod \"certified-operators-xnlzx\" (UID: \"a0309136-eab0-4861-8b0f-b0d9e783bb4c\") " pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.674824 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0309136-eab0-4861-8b0f-b0d9e783bb4c-utilities\") pod \"certified-operators-xnlzx\" (UID: \"a0309136-eab0-4861-8b0f-b0d9e783bb4c\") " pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.776543 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0309136-eab0-4861-8b0f-b0d9e783bb4c-utilities\") pod \"certified-operators-xnlzx\" (UID: \"a0309136-eab0-4861-8b0f-b0d9e783bb4c\") " pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.776648 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9szt\" (UniqueName: \"kubernetes.io/projected/a0309136-eab0-4861-8b0f-b0d9e783bb4c-kube-api-access-r9szt\") pod \"certified-operators-xnlzx\" (UID: \"a0309136-eab0-4861-8b0f-b0d9e783bb4c\") " pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.776700 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0309136-eab0-4861-8b0f-b0d9e783bb4c-catalog-content\") pod \"certified-operators-xnlzx\" (UID: \"a0309136-eab0-4861-8b0f-b0d9e783bb4c\") " pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.777332 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0309136-eab0-4861-8b0f-b0d9e783bb4c-utilities\") pod \"certified-operators-xnlzx\" (UID: \"a0309136-eab0-4861-8b0f-b0d9e783bb4c\") " pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.777397 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0309136-eab0-4861-8b0f-b0d9e783bb4c-catalog-content\") pod \"certified-operators-xnlzx\" (UID: \"a0309136-eab0-4861-8b0f-b0d9e783bb4c\") " pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.802239 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9szt\" (UniqueName: \"kubernetes.io/projected/a0309136-eab0-4861-8b0f-b0d9e783bb4c-kube-api-access-r9szt\") pod \"certified-operators-xnlzx\" (UID: \"a0309136-eab0-4861-8b0f-b0d9e783bb4c\") " pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:49 crc kubenswrapper[4966]: I1217 08:26:49.898028 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.083536 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c9l6f"] Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.085728 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.091313 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db25g" event={"ID":"2cd4b4c7-507c-4f2c-a657-b579d039441b","Type":"ContainerStarted","Data":"10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264"} Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.117778 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c9l6f"] Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.183448 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbj8g\" (UniqueName: \"kubernetes.io/projected/58b62b06-5dd1-46d2-a69b-767595012e59-kube-api-access-mbj8g\") pod \"redhat-operators-c9l6f\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.183656 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-utilities\") pod \"redhat-operators-c9l6f\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.183697 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-catalog-content\") pod \"redhat-operators-c9l6f\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.284826 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbj8g\" (UniqueName: \"kubernetes.io/projected/58b62b06-5dd1-46d2-a69b-767595012e59-kube-api-access-mbj8g\") pod \"redhat-operators-c9l6f\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.284982 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-utilities\") pod \"redhat-operators-c9l6f\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.285012 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-catalog-content\") pod \"redhat-operators-c9l6f\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.285454 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-catalog-content\") pod \"redhat-operators-c9l6f\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.285573 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-utilities\") pod \"redhat-operators-c9l6f\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.304393 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbj8g\" (UniqueName: \"kubernetes.io/projected/58b62b06-5dd1-46d2-a69b-767595012e59-kube-api-access-mbj8g\") pod \"redhat-operators-c9l6f\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.522090 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:26:50 crc kubenswrapper[4966]: I1217 08:26:50.581067 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xnlzx"] Dec 17 08:26:51 crc kubenswrapper[4966]: I1217 08:26:51.097944 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xnlzx" event={"ID":"a0309136-eab0-4861-8b0f-b0d9e783bb4c","Type":"ContainerStarted","Data":"5573fa466aa5bff78c0d390bfdc918e2bdb725d607047f78873f80df7839b9dd"} Dec 17 08:26:51 crc kubenswrapper[4966]: I1217 08:26:51.132574 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c9l6f"] Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.105304 4966 generic.go:334] "Generic (PLEG): container finished" podID="58b62b06-5dd1-46d2-a69b-767595012e59" containerID="7f313097a3fcfdd4729bd70258b4c859b6520cb5075269aa7b390f9d1db418ac" exitCode=0 Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.105387 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9l6f" event={"ID":"58b62b06-5dd1-46d2-a69b-767595012e59","Type":"ContainerDied","Data":"7f313097a3fcfdd4729bd70258b4c859b6520cb5075269aa7b390f9d1db418ac"} Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.105412 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9l6f" event={"ID":"58b62b06-5dd1-46d2-a69b-767595012e59","Type":"ContainerStarted","Data":"a4380601fd215276280c396b9fca9e1127bb4bd273693bf9a046c15515366922"} Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.111179 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nn68p" event={"ID":"4d6d02de-ba82-48f6-aaaf-01e6a65b0dcd","Type":"ContainerStarted","Data":"dd8a90b64c35f380106f36b45e202c1bbd9cc990b84954dc2f5471722822506f"} Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.113184 4966 generic.go:334] "Generic (PLEG): container finished" podID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerID="10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264" exitCode=0 Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.113252 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db25g" event={"ID":"2cd4b4c7-507c-4f2c-a657-b579d039441b","Type":"ContainerDied","Data":"10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264"} Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.117674 4966 generic.go:334] "Generic (PLEG): container finished" podID="a0309136-eab0-4861-8b0f-b0d9e783bb4c" containerID="e11fdf72096ada9747c41fe3362cda2f97943701b042e51b5295a3fe70deb44f" exitCode=0 Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.117716 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xnlzx" event={"ID":"a0309136-eab0-4861-8b0f-b0d9e783bb4c","Type":"ContainerDied","Data":"e11fdf72096ada9747c41fe3362cda2f97943701b042e51b5295a3fe70deb44f"} Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.153408 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nn68p" podStartSLOduration=4.07546549 podStartE2EDuration="7.153390427s" podCreationTimestamp="2025-12-17 08:26:45 +0000 UTC" firstStartedPulling="2025-12-17 08:26:47.031411204 +0000 UTC m=+342.576481146" lastFinishedPulling="2025-12-17 08:26:50.109336141 +0000 UTC m=+345.654406083" observedRunningTime="2025-12-17 08:26:52.152700728 +0000 UTC m=+347.697770670" watchObservedRunningTime="2025-12-17 08:26:52.153390427 +0000 UTC m=+347.698460369" Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.182010 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.182081 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:52 crc kubenswrapper[4966]: I1217 08:26:52.248731 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.216631 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7fjhc" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.758464 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.758525 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.849365 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c678n"] Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.864278 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.875717 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjhpd\" (UniqueName: \"kubernetes.io/projected/7366fd21-5bb2-45f3-b86f-fad3814194a8-kube-api-access-pjhpd\") pod \"certified-operators-c678n\" (UID: \"7366fd21-5bb2-45f3-b86f-fad3814194a8\") " pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.875820 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7366fd21-5bb2-45f3-b86f-fad3814194a8-catalog-content\") pod \"certified-operators-c678n\" (UID: \"7366fd21-5bb2-45f3-b86f-fad3814194a8\") " pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.875935 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7366fd21-5bb2-45f3-b86f-fad3814194a8-utilities\") pod \"certified-operators-c678n\" (UID: \"7366fd21-5bb2-45f3-b86f-fad3814194a8\") " pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.884232 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c678n"] Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.977798 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7366fd21-5bb2-45f3-b86f-fad3814194a8-catalog-content\") pod \"certified-operators-c678n\" (UID: \"7366fd21-5bb2-45f3-b86f-fad3814194a8\") " pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.978525 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7366fd21-5bb2-45f3-b86f-fad3814194a8-catalog-content\") pod \"certified-operators-c678n\" (UID: \"7366fd21-5bb2-45f3-b86f-fad3814194a8\") " pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.978863 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7366fd21-5bb2-45f3-b86f-fad3814194a8-utilities\") pod \"certified-operators-c678n\" (UID: \"7366fd21-5bb2-45f3-b86f-fad3814194a8\") " pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.979111 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjhpd\" (UniqueName: \"kubernetes.io/projected/7366fd21-5bb2-45f3-b86f-fad3814194a8-kube-api-access-pjhpd\") pod \"certified-operators-c678n\" (UID: \"7366fd21-5bb2-45f3-b86f-fad3814194a8\") " pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:53 crc kubenswrapper[4966]: I1217 08:26:53.979389 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7366fd21-5bb2-45f3-b86f-fad3814194a8-utilities\") pod \"certified-operators-c678n\" (UID: \"7366fd21-5bb2-45f3-b86f-fad3814194a8\") " pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:54 crc kubenswrapper[4966]: I1217 08:26:54.008478 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjhpd\" (UniqueName: \"kubernetes.io/projected/7366fd21-5bb2-45f3-b86f-fad3814194a8-kube-api-access-pjhpd\") pod \"certified-operators-c678n\" (UID: \"7366fd21-5bb2-45f3-b86f-fad3814194a8\") " pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:54 crc kubenswrapper[4966]: I1217 08:26:54.137763 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9l6f" event={"ID":"58b62b06-5dd1-46d2-a69b-767595012e59","Type":"ContainerStarted","Data":"c18b05a228a591a133668047e295d663459afc446fb2aa58c0c386ef764235dc"} Dec 17 08:26:54 crc kubenswrapper[4966]: I1217 08:26:54.139620 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db25g" event={"ID":"2cd4b4c7-507c-4f2c-a657-b579d039441b","Type":"ContainerStarted","Data":"09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f"} Dec 17 08:26:54 crc kubenswrapper[4966]: I1217 08:26:54.141751 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xnlzx" event={"ID":"a0309136-eab0-4861-8b0f-b0d9e783bb4c","Type":"ContainerStarted","Data":"39051d1fa98d6921ae597e43a2d93c970386e44603654f91574e0e9cf119625c"} Dec 17 08:26:54 crc kubenswrapper[4966]: I1217 08:26:54.191950 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:26:54 crc kubenswrapper[4966]: I1217 08:26:54.195588 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-db25g" podStartSLOduration=3.085963886 podStartE2EDuration="8.195570449s" podCreationTimestamp="2025-12-17 08:26:46 +0000 UTC" firstStartedPulling="2025-12-17 08:26:48.034776034 +0000 UTC m=+343.579845976" lastFinishedPulling="2025-12-17 08:26:53.144382587 +0000 UTC m=+348.689452539" observedRunningTime="2025-12-17 08:26:54.19101072 +0000 UTC m=+349.736080662" watchObservedRunningTime="2025-12-17 08:26:54.195570449 +0000 UTC m=+349.740640391" Dec 17 08:26:54 crc kubenswrapper[4966]: I1217 08:26:54.823976 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pv2kq" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="registry-server" probeResult="failure" output=< Dec 17 08:26:54 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:26:54 crc kubenswrapper[4966]: > Dec 17 08:26:54 crc kubenswrapper[4966]: I1217 08:26:54.856478 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c678n"] Dec 17 08:26:55 crc kubenswrapper[4966]: I1217 08:26:55.174548 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c678n" event={"ID":"7366fd21-5bb2-45f3-b86f-fad3814194a8","Type":"ContainerStarted","Data":"1d7b839b9f26b2a1eb4985cd5b3fd4da56e111feb6763dc6c30dc9ae684f66c6"} Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.161497 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.161797 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.185249 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c678n" event={"ID":"7366fd21-5bb2-45f3-b86f-fad3814194a8","Type":"ContainerStarted","Data":"a3bea3dd3aaa2c76d610bb27bac4b99506617e275ccc803a575554ad8039475f"} Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.477983 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.533702 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nn68p" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.644763 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6j8ws"] Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.646174 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.649205 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.659168 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6j8ws"] Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.739990 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a6a0b1-0281-449e-8c54-3963ccc8e098-utilities\") pod \"community-operators-6j8ws\" (UID: \"75a6a0b1-0281-449e-8c54-3963ccc8e098\") " pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.740320 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a6a0b1-0281-449e-8c54-3963ccc8e098-catalog-content\") pod \"community-operators-6j8ws\" (UID: \"75a6a0b1-0281-449e-8c54-3963ccc8e098\") " pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.740474 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcr96\" (UniqueName: \"kubernetes.io/projected/75a6a0b1-0281-449e-8c54-3963ccc8e098-kube-api-access-tcr96\") pod \"community-operators-6j8ws\" (UID: \"75a6a0b1-0281-449e-8c54-3963ccc8e098\") " pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.768382 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.768910 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.842369 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a6a0b1-0281-449e-8c54-3963ccc8e098-utilities\") pod \"community-operators-6j8ws\" (UID: \"75a6a0b1-0281-449e-8c54-3963ccc8e098\") " pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.843196 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a6a0b1-0281-449e-8c54-3963ccc8e098-catalog-content\") pod \"community-operators-6j8ws\" (UID: \"75a6a0b1-0281-449e-8c54-3963ccc8e098\") " pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.843326 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcr96\" (UniqueName: \"kubernetes.io/projected/75a6a0b1-0281-449e-8c54-3963ccc8e098-kube-api-access-tcr96\") pod \"community-operators-6j8ws\" (UID: \"75a6a0b1-0281-449e-8c54-3963ccc8e098\") " pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.844578 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a6a0b1-0281-449e-8c54-3963ccc8e098-catalog-content\") pod \"community-operators-6j8ws\" (UID: \"75a6a0b1-0281-449e-8c54-3963ccc8e098\") " pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.846438 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a6a0b1-0281-449e-8c54-3963ccc8e098-utilities\") pod \"community-operators-6j8ws\" (UID: \"75a6a0b1-0281-449e-8c54-3963ccc8e098\") " pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.867029 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcr96\" (UniqueName: \"kubernetes.io/projected/75a6a0b1-0281-449e-8c54-3963ccc8e098-kube-api-access-tcr96\") pod \"community-operators-6j8ws\" (UID: \"75a6a0b1-0281-449e-8c54-3963ccc8e098\") " pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:56 crc kubenswrapper[4966]: I1217 08:26:56.960762 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:26:57 crc kubenswrapper[4966]: I1217 08:26:57.400334 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6j8ws"] Dec 17 08:26:57 crc kubenswrapper[4966]: I1217 08:26:57.824367 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-db25g" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerName="registry-server" probeResult="failure" output=< Dec 17 08:26:57 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:26:57 crc kubenswrapper[4966]: > Dec 17 08:26:58 crc kubenswrapper[4966]: I1217 08:26:58.197320 4966 generic.go:334] "Generic (PLEG): container finished" podID="a0309136-eab0-4861-8b0f-b0d9e783bb4c" containerID="39051d1fa98d6921ae597e43a2d93c970386e44603654f91574e0e9cf119625c" exitCode=0 Dec 17 08:26:58 crc kubenswrapper[4966]: I1217 08:26:58.197401 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xnlzx" event={"ID":"a0309136-eab0-4861-8b0f-b0d9e783bb4c","Type":"ContainerDied","Data":"39051d1fa98d6921ae597e43a2d93c970386e44603654f91574e0e9cf119625c"} Dec 17 08:26:58 crc kubenswrapper[4966]: I1217 08:26:58.198402 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6j8ws" event={"ID":"75a6a0b1-0281-449e-8c54-3963ccc8e098","Type":"ContainerStarted","Data":"e93ebadcafe201df4566ef2f50609362101733213552bcdd6e70c83b2e392b6c"} Dec 17 08:26:59 crc kubenswrapper[4966]: I1217 08:26:59.204737 4966 generic.go:334] "Generic (PLEG): container finished" podID="75a6a0b1-0281-449e-8c54-3963ccc8e098" containerID="ae4010b3df5cc2cb12c78dd54cc395e13f523687f436f509ab9e26eedaacbc14" exitCode=0 Dec 17 08:26:59 crc kubenswrapper[4966]: I1217 08:26:59.204805 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6j8ws" event={"ID":"75a6a0b1-0281-449e-8c54-3963ccc8e098","Type":"ContainerDied","Data":"ae4010b3df5cc2cb12c78dd54cc395e13f523687f436f509ab9e26eedaacbc14"} Dec 17 08:27:00 crc kubenswrapper[4966]: I1217 08:27:00.212432 4966 generic.go:334] "Generic (PLEG): container finished" podID="58b62b06-5dd1-46d2-a69b-767595012e59" containerID="c18b05a228a591a133668047e295d663459afc446fb2aa58c0c386ef764235dc" exitCode=0 Dec 17 08:27:00 crc kubenswrapper[4966]: I1217 08:27:00.212513 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9l6f" event={"ID":"58b62b06-5dd1-46d2-a69b-767595012e59","Type":"ContainerDied","Data":"c18b05a228a591a133668047e295d663459afc446fb2aa58c0c386ef764235dc"} Dec 17 08:27:00 crc kubenswrapper[4966]: I1217 08:27:00.215975 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xnlzx" event={"ID":"a0309136-eab0-4861-8b0f-b0d9e783bb4c","Type":"ContainerStarted","Data":"d99627a308fd7951c565fbeb47bbb51533e4f7d4551336add776a856fac6be5c"} Dec 17 08:27:00 crc kubenswrapper[4966]: I1217 08:27:00.219098 4966 generic.go:334] "Generic (PLEG): container finished" podID="7366fd21-5bb2-45f3-b86f-fad3814194a8" containerID="a3bea3dd3aaa2c76d610bb27bac4b99506617e275ccc803a575554ad8039475f" exitCode=0 Dec 17 08:27:00 crc kubenswrapper[4966]: I1217 08:27:00.219613 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c678n" event={"ID":"7366fd21-5bb2-45f3-b86f-fad3814194a8","Type":"ContainerDied","Data":"a3bea3dd3aaa2c76d610bb27bac4b99506617e275ccc803a575554ad8039475f"} Dec 17 08:27:00 crc kubenswrapper[4966]: I1217 08:27:00.273548 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xnlzx" podStartSLOduration=3.430720346 podStartE2EDuration="11.273532955s" podCreationTimestamp="2025-12-17 08:26:49 +0000 UTC" firstStartedPulling="2025-12-17 08:26:52.119317114 +0000 UTC m=+347.664387056" lastFinishedPulling="2025-12-17 08:26:59.962129723 +0000 UTC m=+355.507199665" observedRunningTime="2025-12-17 08:27:00.269419189 +0000 UTC m=+355.814489131" watchObservedRunningTime="2025-12-17 08:27:00.273532955 +0000 UTC m=+355.818602897" Dec 17 08:27:03 crc kubenswrapper[4966]: I1217 08:27:03.805556 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:27:03 crc kubenswrapper[4966]: I1217 08:27:03.841908 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 08:27:06 crc kubenswrapper[4966]: I1217 08:27:06.264925 4966 generic.go:334] "Generic (PLEG): container finished" podID="7366fd21-5bb2-45f3-b86f-fad3814194a8" containerID="f5de8d011e2beea89467ac5a0508f0285d684c49b4581ed6ffd5959fe8a2b98d" exitCode=0 Dec 17 08:27:06 crc kubenswrapper[4966]: I1217 08:27:06.264991 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c678n" event={"ID":"7366fd21-5bb2-45f3-b86f-fad3814194a8","Type":"ContainerDied","Data":"f5de8d011e2beea89467ac5a0508f0285d684c49b4581ed6ffd5959fe8a2b98d"} Dec 17 08:27:06 crc kubenswrapper[4966]: I1217 08:27:06.268753 4966 generic.go:334] "Generic (PLEG): container finished" podID="75a6a0b1-0281-449e-8c54-3963ccc8e098" containerID="38c9555836b0e207c9793b3270d8219b339a5e3a5bee02bcdd3278698eb4ecb0" exitCode=0 Dec 17 08:27:06 crc kubenswrapper[4966]: I1217 08:27:06.268802 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6j8ws" event={"ID":"75a6a0b1-0281-449e-8c54-3963ccc8e098","Type":"ContainerDied","Data":"38c9555836b0e207c9793b3270d8219b339a5e3a5bee02bcdd3278698eb4ecb0"} Dec 17 08:27:06 crc kubenswrapper[4966]: I1217 08:27:06.270886 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9l6f" event={"ID":"58b62b06-5dd1-46d2-a69b-767595012e59","Type":"ContainerStarted","Data":"9af75d3f8f4a710b951f69ce7a34f48209b303ed6a34b72a208f64e3bb82577b"} Dec 17 08:27:06 crc kubenswrapper[4966]: I1217 08:27:06.334447 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c9l6f" podStartSLOduration=3.199863725 podStartE2EDuration="16.334430087s" podCreationTimestamp="2025-12-17 08:26:50 +0000 UTC" firstStartedPulling="2025-12-17 08:26:52.107334105 +0000 UTC m=+347.652404047" lastFinishedPulling="2025-12-17 08:27:05.241900467 +0000 UTC m=+360.786970409" observedRunningTime="2025-12-17 08:27:06.331549325 +0000 UTC m=+361.876619277" watchObservedRunningTime="2025-12-17 08:27:06.334430087 +0000 UTC m=+361.879500019" Dec 17 08:27:06 crc kubenswrapper[4966]: I1217 08:27:06.844899 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:27:06 crc kubenswrapper[4966]: I1217 08:27:06.928216 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-db25g" Dec 17 08:27:07 crc kubenswrapper[4966]: I1217 08:27:07.277349 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6j8ws" event={"ID":"75a6a0b1-0281-449e-8c54-3963ccc8e098","Type":"ContainerStarted","Data":"d8f343d323f9f6fd2fcafb43575df8691fea5a22e7311bf19b4be7908196c1fa"} Dec 17 08:27:07 crc kubenswrapper[4966]: I1217 08:27:07.280357 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c678n" event={"ID":"7366fd21-5bb2-45f3-b86f-fad3814194a8","Type":"ContainerStarted","Data":"1f96c271e09f62e760f5edb70a11a07b0a4288176b404481272e13ee1b38dfef"} Dec 17 08:27:07 crc kubenswrapper[4966]: I1217 08:27:07.302269 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6j8ws" podStartSLOduration=4.767656992 podStartE2EDuration="11.302248223s" podCreationTimestamp="2025-12-17 08:26:56 +0000 UTC" firstStartedPulling="2025-12-17 08:27:00.220941458 +0000 UTC m=+355.766011410" lastFinishedPulling="2025-12-17 08:27:06.755532699 +0000 UTC m=+362.300602641" observedRunningTime="2025-12-17 08:27:07.299773823 +0000 UTC m=+362.844843765" watchObservedRunningTime="2025-12-17 08:27:07.302248223 +0000 UTC m=+362.847318175" Dec 17 08:27:07 crc kubenswrapper[4966]: I1217 08:27:07.322603 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c678n" podStartSLOduration=7.877958288 podStartE2EDuration="14.322585037s" podCreationTimestamp="2025-12-17 08:26:53 +0000 UTC" firstStartedPulling="2025-12-17 08:27:00.235064287 +0000 UTC m=+355.780134249" lastFinishedPulling="2025-12-17 08:27:06.679691056 +0000 UTC m=+362.224760998" observedRunningTime="2025-12-17 08:27:07.320846158 +0000 UTC m=+362.865916110" watchObservedRunningTime="2025-12-17 08:27:07.322585037 +0000 UTC m=+362.867654979" Dec 17 08:27:09 crc kubenswrapper[4966]: I1217 08:27:09.898234 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:27:09 crc kubenswrapper[4966]: I1217 08:27:09.898313 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:27:10 crc kubenswrapper[4966]: I1217 08:27:10.069982 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:27:10 crc kubenswrapper[4966]: I1217 08:27:10.384700 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xnlzx" Dec 17 08:27:10 crc kubenswrapper[4966]: I1217 08:27:10.522961 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:27:10 crc kubenswrapper[4966]: I1217 08:27:10.523023 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:27:11 crc kubenswrapper[4966]: I1217 08:27:11.669628 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c9l6f" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" containerName="registry-server" probeResult="failure" output=< Dec 17 08:27:11 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:27:11 crc kubenswrapper[4966]: > Dec 17 08:27:14 crc kubenswrapper[4966]: I1217 08:27:14.196365 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:27:14 crc kubenswrapper[4966]: I1217 08:27:14.196432 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:27:14 crc kubenswrapper[4966]: I1217 08:27:14.238491 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:27:14 crc kubenswrapper[4966]: I1217 08:27:14.438358 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c678n" Dec 17 08:27:16 crc kubenswrapper[4966]: I1217 08:27:16.807664 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:27:16 crc kubenswrapper[4966]: I1217 08:27:16.807735 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:27:16 crc kubenswrapper[4966]: I1217 08:27:16.961574 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:27:16 crc kubenswrapper[4966]: I1217 08:27:16.961649 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:27:17 crc kubenswrapper[4966]: I1217 08:27:17.002481 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:27:17 crc kubenswrapper[4966]: I1217 08:27:17.447501 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6j8ws" Dec 17 08:27:20 crc kubenswrapper[4966]: I1217 08:27:20.585735 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:27:20 crc kubenswrapper[4966]: I1217 08:27:20.643802 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.738622 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x8k4s"] Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.739537 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.750608 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x8k4s"] Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.866503 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2bp7\" (UniqueName: \"kubernetes.io/projected/5ef61f83-2f18-4fa7-816d-1feafe75570c-kube-api-access-m2bp7\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.866597 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.866636 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ef61f83-2f18-4fa7-816d-1feafe75570c-bound-sa-token\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.866654 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ef61f83-2f18-4fa7-816d-1feafe75570c-registry-tls\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.866703 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ef61f83-2f18-4fa7-816d-1feafe75570c-trusted-ca\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.866760 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ef61f83-2f18-4fa7-816d-1feafe75570c-registry-certificates\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.866794 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ef61f83-2f18-4fa7-816d-1feafe75570c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.866813 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ef61f83-2f18-4fa7-816d-1feafe75570c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.887601 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.967818 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ef61f83-2f18-4fa7-816d-1feafe75570c-bound-sa-token\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.967871 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ef61f83-2f18-4fa7-816d-1feafe75570c-registry-tls\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.967941 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ef61f83-2f18-4fa7-816d-1feafe75570c-trusted-ca\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.967981 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ef61f83-2f18-4fa7-816d-1feafe75570c-registry-certificates\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.968016 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ef61f83-2f18-4fa7-816d-1feafe75570c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.968033 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ef61f83-2f18-4fa7-816d-1feafe75570c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.968052 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2bp7\" (UniqueName: \"kubernetes.io/projected/5ef61f83-2f18-4fa7-816d-1feafe75570c-kube-api-access-m2bp7\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.968766 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5ef61f83-2f18-4fa7-816d-1feafe75570c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.969199 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ef61f83-2f18-4fa7-816d-1feafe75570c-trusted-ca\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.969203 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5ef61f83-2f18-4fa7-816d-1feafe75570c-registry-certificates\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.974035 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5ef61f83-2f18-4fa7-816d-1feafe75570c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.978069 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5ef61f83-2f18-4fa7-816d-1feafe75570c-registry-tls\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:21 crc kubenswrapper[4966]: I1217 08:27:21.986316 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ef61f83-2f18-4fa7-816d-1feafe75570c-bound-sa-token\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:22 crc kubenswrapper[4966]: I1217 08:27:22.002432 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2bp7\" (UniqueName: \"kubernetes.io/projected/5ef61f83-2f18-4fa7-816d-1feafe75570c-kube-api-access-m2bp7\") pod \"image-registry-66df7c8f76-x8k4s\" (UID: \"5ef61f83-2f18-4fa7-816d-1feafe75570c\") " pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:22 crc kubenswrapper[4966]: I1217 08:27:22.055164 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:22 crc kubenswrapper[4966]: I1217 08:27:22.518051 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x8k4s"] Dec 17 08:27:22 crc kubenswrapper[4966]: W1217 08:27:22.519658 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ef61f83_2f18_4fa7_816d_1feafe75570c.slice/crio-fa9d2572e3e4941bda74d64b99f72f1346b17ccf090d91ed8ab2e9333c4d21b5 WatchSource:0}: Error finding container fa9d2572e3e4941bda74d64b99f72f1346b17ccf090d91ed8ab2e9333c4d21b5: Status 404 returned error can't find the container with id fa9d2572e3e4941bda74d64b99f72f1346b17ccf090d91ed8ab2e9333c4d21b5 Dec 17 08:27:23 crc kubenswrapper[4966]: I1217 08:27:23.435054 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" event={"ID":"5ef61f83-2f18-4fa7-816d-1feafe75570c","Type":"ContainerStarted","Data":"380bdd1b4e6e65c6eb3930945d3d75ca959501efd93df10c38753d15ba093cb9"} Dec 17 08:27:23 crc kubenswrapper[4966]: I1217 08:27:23.435371 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:23 crc kubenswrapper[4966]: I1217 08:27:23.435382 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" event={"ID":"5ef61f83-2f18-4fa7-816d-1feafe75570c","Type":"ContainerStarted","Data":"fa9d2572e3e4941bda74d64b99f72f1346b17ccf090d91ed8ab2e9333c4d21b5"} Dec 17 08:27:23 crc kubenswrapper[4966]: I1217 08:27:23.546132 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" podStartSLOduration=2.546113617 podStartE2EDuration="2.546113617s" podCreationTimestamp="2025-12-17 08:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:27:23.541593649 +0000 UTC m=+379.086663591" watchObservedRunningTime="2025-12-17 08:27:23.546113617 +0000 UTC m=+379.091183559" Dec 17 08:27:35 crc kubenswrapper[4966]: I1217 08:27:35.920440 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:27:35 crc kubenswrapper[4966]: I1217 08:27:35.920897 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:27:35 crc kubenswrapper[4966]: I1217 08:27:35.923220 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:27:35 crc kubenswrapper[4966]: I1217 08:27:35.935540 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:27:36 crc kubenswrapper[4966]: I1217 08:27:36.031443 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 17 08:27:36 crc kubenswrapper[4966]: I1217 08:27:36.507221 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"61effef45178db06d615072abd6e769d25e324f950e1391dee007f2211954d3b"} Dec 17 08:27:36 crc kubenswrapper[4966]: I1217 08:27:36.507519 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"059a1475139b78e983050a011a9b707b69f579ab23800e52facd951ed747510d"} Dec 17 08:27:36 crc kubenswrapper[4966]: I1217 08:27:36.833333 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:27:36 crc kubenswrapper[4966]: I1217 08:27:36.833636 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:27:36 crc kubenswrapper[4966]: I1217 08:27:36.840432 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:27:36 crc kubenswrapper[4966]: I1217 08:27:36.840469 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:27:36 crc kubenswrapper[4966]: I1217 08:27:36.931588 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 17 08:27:37 crc kubenswrapper[4966]: W1217 08:27:37.130657 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-727aa1fb9dbec943d6db123de47e0b59e1b3180277a9cef5c5f239f5a2ea7859 WatchSource:0}: Error finding container 727aa1fb9dbec943d6db123de47e0b59e1b3180277a9cef5c5f239f5a2ea7859: Status 404 returned error can't find the container with id 727aa1fb9dbec943d6db123de47e0b59e1b3180277a9cef5c5f239f5a2ea7859 Dec 17 08:27:37 crc kubenswrapper[4966]: I1217 08:27:37.131068 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:27:37 crc kubenswrapper[4966]: I1217 08:27:37.514236 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2945c568357bf9ea5af131cdcd08801302f40aaa80a933fddbdf805abdfc9ab1"} Dec 17 08:27:37 crc kubenswrapper[4966]: I1217 08:27:37.514777 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"727aa1fb9dbec943d6db123de47e0b59e1b3180277a9cef5c5f239f5a2ea7859"} Dec 17 08:27:38 crc kubenswrapper[4966]: I1217 08:27:38.519369 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"71a2d94373e63b91a255403ff2919d059380848a0486b6702ea3544b91acd900"} Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.294268 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-64585bb48f-stzs9"] Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.294744 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" podUID="42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" containerName="controller-manager" containerID="cri-o://1a0c4b4f11c39c6d66941a957903608f3dbc7f965cc926cc62e99e68e571a31f" gracePeriod=30 Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.531963 4966 generic.go:334] "Generic (PLEG): container finished" podID="42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" containerID="1a0c4b4f11c39c6d66941a957903608f3dbc7f965cc926cc62e99e68e571a31f" exitCode=0 Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.532014 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" event={"ID":"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2","Type":"ContainerDied","Data":"1a0c4b4f11c39c6d66941a957903608f3dbc7f965cc926cc62e99e68e571a31f"} Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.535935 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"989e1676a24f996c33a1bff43c351fe186a8605e8c64bcf12d9a951eeb307ac6"} Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.537006 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.659908 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.825531 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bw2k\" (UniqueName: \"kubernetes.io/projected/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-kube-api-access-6bw2k\") pod \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.825594 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-proxy-ca-bundles\") pod \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.825665 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-serving-cert\") pod \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.825706 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-config\") pod \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.825736 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-client-ca\") pod \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\" (UID: \"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2\") " Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.826460 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-config" (OuterVolumeSpecName: "config") pod "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" (UID: "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.826481 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-client-ca" (OuterVolumeSpecName: "client-ca") pod "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" (UID: "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.826892 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" (UID: "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.831083 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" (UID: "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.833005 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-kube-api-access-6bw2k" (OuterVolumeSpecName: "kube-api-access-6bw2k") pod "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" (UID: "42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2"). InnerVolumeSpecName "kube-api-access-6bw2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.927241 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bw2k\" (UniqueName: \"kubernetes.io/projected/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-kube-api-access-6bw2k\") on node \"crc\" DevicePath \"\"" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.927281 4966 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.927293 4966 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.927306 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:27:39 crc kubenswrapper[4966]: I1217 08:27:39.927330 4966 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2-client-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.543093 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" event={"ID":"42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2","Type":"ContainerDied","Data":"9536e7cb7e84fb58ecf8cbb50cc1ace9b57f218932fdfa39e1eb254e28e457b3"} Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.543137 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64585bb48f-stzs9" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.543141 4966 scope.go:117] "RemoveContainer" containerID="1a0c4b4f11c39c6d66941a957903608f3dbc7f965cc926cc62e99e68e571a31f" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.576659 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-64585bb48f-stzs9"] Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.584725 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-64585bb48f-stzs9"] Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.667890 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b46f89db7-6lmgj"] Dec 17 08:27:40 crc kubenswrapper[4966]: E1217 08:27:40.668129 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" containerName="controller-manager" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.668146 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" containerName="controller-manager" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.668264 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" containerName="controller-manager" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.668720 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.670914 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.671033 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.671838 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.673026 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.673179 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.674093 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.678682 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b46f89db7-6lmgj"] Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.686225 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.839063 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d446236-5069-43fc-bf10-2f1f4353a503-config\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.839132 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2" path="/var/lib/kubelet/pods/42ac0063-2a1f-4890-b4b5-ae34bf3c3bf2/volumes" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.839174 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5d446236-5069-43fc-bf10-2f1f4353a503-proxy-ca-bundles\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.839268 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d446236-5069-43fc-bf10-2f1f4353a503-serving-cert\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.839309 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d446236-5069-43fc-bf10-2f1f4353a503-client-ca\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.839369 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8jxt\" (UniqueName: \"kubernetes.io/projected/5d446236-5069-43fc-bf10-2f1f4353a503-kube-api-access-k8jxt\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.940399 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5d446236-5069-43fc-bf10-2f1f4353a503-proxy-ca-bundles\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.940476 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d446236-5069-43fc-bf10-2f1f4353a503-serving-cert\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.940508 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d446236-5069-43fc-bf10-2f1f4353a503-client-ca\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.940534 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8jxt\" (UniqueName: \"kubernetes.io/projected/5d446236-5069-43fc-bf10-2f1f4353a503-kube-api-access-k8jxt\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.940564 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d446236-5069-43fc-bf10-2f1f4353a503-config\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.941851 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d446236-5069-43fc-bf10-2f1f4353a503-config\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.942732 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5d446236-5069-43fc-bf10-2f1f4353a503-proxy-ca-bundles\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.944354 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d446236-5069-43fc-bf10-2f1f4353a503-client-ca\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.956381 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d446236-5069-43fc-bf10-2f1f4353a503-serving-cert\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.964152 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8jxt\" (UniqueName: \"kubernetes.io/projected/5d446236-5069-43fc-bf10-2f1f4353a503-kube-api-access-k8jxt\") pod \"controller-manager-5b46f89db7-6lmgj\" (UID: \"5d446236-5069-43fc-bf10-2f1f4353a503\") " pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:40 crc kubenswrapper[4966]: I1217 08:27:40.985950 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:41 crc kubenswrapper[4966]: I1217 08:27:41.173604 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b46f89db7-6lmgj"] Dec 17 08:27:41 crc kubenswrapper[4966]: W1217 08:27:41.183851 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d446236_5069_43fc_bf10_2f1f4353a503.slice/crio-0c37ebdd1887c9fc4293841e7c77ddb1d2f58fdf894b4ca02c83488f875dc8aa WatchSource:0}: Error finding container 0c37ebdd1887c9fc4293841e7c77ddb1d2f58fdf894b4ca02c83488f875dc8aa: Status 404 returned error can't find the container with id 0c37ebdd1887c9fc4293841e7c77ddb1d2f58fdf894b4ca02c83488f875dc8aa Dec 17 08:27:41 crc kubenswrapper[4966]: I1217 08:27:41.550432 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" event={"ID":"5d446236-5069-43fc-bf10-2f1f4353a503","Type":"ContainerStarted","Data":"ed8207977926d92b9ede8ac4b56dcbcecf5f1a979a5d9763775c5acd44669d94"} Dec 17 08:27:41 crc kubenswrapper[4966]: I1217 08:27:41.550756 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" event={"ID":"5d446236-5069-43fc-bf10-2f1f4353a503","Type":"ContainerStarted","Data":"0c37ebdd1887c9fc4293841e7c77ddb1d2f58fdf894b4ca02c83488f875dc8aa"} Dec 17 08:27:41 crc kubenswrapper[4966]: I1217 08:27:41.584931 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" podStartSLOduration=2.58491401 podStartE2EDuration="2.58491401s" podCreationTimestamp="2025-12-17 08:27:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:27:41.581206425 +0000 UTC m=+397.126276367" watchObservedRunningTime="2025-12-17 08:27:41.58491401 +0000 UTC m=+397.129983952" Dec 17 08:27:42 crc kubenswrapper[4966]: I1217 08:27:42.068388 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-x8k4s" Dec 17 08:27:42 crc kubenswrapper[4966]: I1217 08:27:42.132231 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sfbnp"] Dec 17 08:27:42 crc kubenswrapper[4966]: I1217 08:27:42.556090 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:42 crc kubenswrapper[4966]: I1217 08:27:42.560752 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b46f89db7-6lmgj" Dec 17 08:27:46 crc kubenswrapper[4966]: I1217 08:27:46.807227 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:27:46 crc kubenswrapper[4966]: I1217 08:27:46.807561 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.184860 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" podUID="d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" containerName="registry" containerID="cri-o://d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2" gracePeriod=30 Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.693916 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.701108 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-bound-sa-token\") pod \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.701157 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-certificates\") pod \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.701202 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-tls\") pod \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.701234 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-trusted-ca\") pod \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.701256 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvjnc\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-kube-api-access-lvjnc\") pod \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.701444 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.701471 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-ca-trust-extracted\") pod \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.701513 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-installation-pull-secrets\") pod \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\" (UID: \"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a\") " Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.702630 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.702794 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.708263 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.712248 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-kube-api-access-lvjnc" (OuterVolumeSpecName: "kube-api-access-lvjnc") pod "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a"). InnerVolumeSpecName "kube-api-access-lvjnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.715739 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.716326 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.717162 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.733746 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" (UID: "d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.735887 4966 generic.go:334] "Generic (PLEG): container finished" podID="d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" containerID="d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2" exitCode=0 Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.735937 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" event={"ID":"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a","Type":"ContainerDied","Data":"d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2"} Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.735974 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" event={"ID":"d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a","Type":"ContainerDied","Data":"92c1092c1be301494b7b99760bd966c0d4cb19dd3c92dbffc42b4e552a8fcc4c"} Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.735998 4966 scope.go:117] "RemoveContainer" containerID="d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.736154 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sfbnp" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.790965 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sfbnp"] Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.792019 4966 scope.go:117] "RemoveContainer" containerID="d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.794763 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sfbnp"] Dec 17 08:28:07 crc kubenswrapper[4966]: E1217 08:28:07.797253 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2\": container with ID starting with d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2 not found: ID does not exist" containerID="d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.797327 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2"} err="failed to get container status \"d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2\": rpc error: code = NotFound desc = could not find container \"d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2\": container with ID starting with d8d6b09e9f3065607c4fe151138b74672f2f29ce9cd653be617cb20cc12c2bc2 not found: ID does not exist" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.802627 4966 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.802662 4966 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.802675 4966 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.802694 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.802735 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvjnc\" (UniqueName: \"kubernetes.io/projected/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-kube-api-access-lvjnc\") on node \"crc\" DevicePath \"\"" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.802747 4966 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 17 08:28:07 crc kubenswrapper[4966]: I1217 08:28:07.802760 4966 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 17 08:28:08 crc kubenswrapper[4966]: I1217 08:28:08.842865 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" path="/var/lib/kubelet/pods/d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a/volumes" Dec 17 08:28:16 crc kubenswrapper[4966]: I1217 08:28:16.807910 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:28:16 crc kubenswrapper[4966]: I1217 08:28:16.808411 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:28:16 crc kubenswrapper[4966]: I1217 08:28:16.808469 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:28:16 crc kubenswrapper[4966]: I1217 08:28:16.809298 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af9614c017f65ecd41597b9880f18912c6650c9ea6e2d776d690cac78b26db56"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 08:28:16 crc kubenswrapper[4966]: I1217 08:28:16.809358 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://af9614c017f65ecd41597b9880f18912c6650c9ea6e2d776d690cac78b26db56" gracePeriod=600 Dec 17 08:28:17 crc kubenswrapper[4966]: I1217 08:28:17.135986 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 17 08:28:17 crc kubenswrapper[4966]: I1217 08:28:17.826736 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="af9614c017f65ecd41597b9880f18912c6650c9ea6e2d776d690cac78b26db56" exitCode=0 Dec 17 08:28:17 crc kubenswrapper[4966]: I1217 08:28:17.826780 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"af9614c017f65ecd41597b9880f18912c6650c9ea6e2d776d690cac78b26db56"} Dec 17 08:28:17 crc kubenswrapper[4966]: I1217 08:28:17.826830 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"ee72e03cd88492b874cbcd116164fd7756b1677dac83cef997b128cc487a5628"} Dec 17 08:28:17 crc kubenswrapper[4966]: I1217 08:28:17.826847 4966 scope.go:117] "RemoveContainer" containerID="21411bc2d6c8605c18f014f1622be4f5bc1e57202a8173b2a36bf1c74cfac532" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.204537 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw"] Dec 17 08:30:00 crc kubenswrapper[4966]: E1217 08:30:00.205738 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" containerName="registry" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.205771 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" containerName="registry" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.206084 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d292a6d1-4821-4ebe-8e64-b5ea9fe4cd1a" containerName="registry" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.207060 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.210900 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.212790 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw"] Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.214153 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.326380 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-secret-volume\") pod \"collect-profiles-29432670-wx7lw\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.326793 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swn8g\" (UniqueName: \"kubernetes.io/projected/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-kube-api-access-swn8g\") pod \"collect-profiles-29432670-wx7lw\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.326839 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-config-volume\") pod \"collect-profiles-29432670-wx7lw\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.428110 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-config-volume\") pod \"collect-profiles-29432670-wx7lw\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.428372 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-secret-volume\") pod \"collect-profiles-29432670-wx7lw\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.428474 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swn8g\" (UniqueName: \"kubernetes.io/projected/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-kube-api-access-swn8g\") pod \"collect-profiles-29432670-wx7lw\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.429690 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-config-volume\") pod \"collect-profiles-29432670-wx7lw\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.438333 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-secret-volume\") pod \"collect-profiles-29432670-wx7lw\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.450975 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swn8g\" (UniqueName: \"kubernetes.io/projected/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-kube-api-access-swn8g\") pod \"collect-profiles-29432670-wx7lw\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.536577 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:00 crc kubenswrapper[4966]: I1217 08:30:00.763300 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw"] Dec 17 08:30:00 crc kubenswrapper[4966]: W1217 08:30:00.779807 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46fa2d36_361a_44c2_b3dc_e8dc0b6d80e2.slice/crio-9d72077cf96f31ed1f10abd7fc23c016d8fbd1c65c73a6128d44fc5f64c9f765 WatchSource:0}: Error finding container 9d72077cf96f31ed1f10abd7fc23c016d8fbd1c65c73a6128d44fc5f64c9f765: Status 404 returned error can't find the container with id 9d72077cf96f31ed1f10abd7fc23c016d8fbd1c65c73a6128d44fc5f64c9f765 Dec 17 08:30:01 crc kubenswrapper[4966]: E1217 08:30:01.090261 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46fa2d36_361a_44c2_b3dc_e8dc0b6d80e2.slice/crio-90516996cc28c3c17310dc503dd6ff471ebabc6871623945ce7800777f33d954.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46fa2d36_361a_44c2_b3dc_e8dc0b6d80e2.slice/crio-conmon-90516996cc28c3c17310dc503dd6ff471ebabc6871623945ce7800777f33d954.scope\": RecentStats: unable to find data in memory cache]" Dec 17 08:30:01 crc kubenswrapper[4966]: I1217 08:30:01.476150 4966 generic.go:334] "Generic (PLEG): container finished" podID="46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2" containerID="90516996cc28c3c17310dc503dd6ff471ebabc6871623945ce7800777f33d954" exitCode=0 Dec 17 08:30:01 crc kubenswrapper[4966]: I1217 08:30:01.476194 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" event={"ID":"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2","Type":"ContainerDied","Data":"90516996cc28c3c17310dc503dd6ff471ebabc6871623945ce7800777f33d954"} Dec 17 08:30:01 crc kubenswrapper[4966]: I1217 08:30:01.476218 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" event={"ID":"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2","Type":"ContainerStarted","Data":"9d72077cf96f31ed1f10abd7fc23c016d8fbd1c65c73a6128d44fc5f64c9f765"} Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.740205 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.858150 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-secret-volume\") pod \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.858270 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swn8g\" (UniqueName: \"kubernetes.io/projected/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-kube-api-access-swn8g\") pod \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.858325 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-config-volume\") pod \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\" (UID: \"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2\") " Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.859393 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-config-volume" (OuterVolumeSpecName: "config-volume") pod "46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2" (UID: "46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.863808 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-kube-api-access-swn8g" (OuterVolumeSpecName: "kube-api-access-swn8g") pod "46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2" (UID: "46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2"). InnerVolumeSpecName "kube-api-access-swn8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.864116 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2" (UID: "46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.960454 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swn8g\" (UniqueName: \"kubernetes.io/projected/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-kube-api-access-swn8g\") on node \"crc\" DevicePath \"\"" Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.960494 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 08:30:02 crc kubenswrapper[4966]: I1217 08:30:02.960510 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 08:30:03 crc kubenswrapper[4966]: I1217 08:30:03.491536 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" event={"ID":"46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2","Type":"ContainerDied","Data":"9d72077cf96f31ed1f10abd7fc23c016d8fbd1c65c73a6128d44fc5f64c9f765"} Dec 17 08:30:03 crc kubenswrapper[4966]: I1217 08:30:03.491580 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw" Dec 17 08:30:03 crc kubenswrapper[4966]: I1217 08:30:03.491587 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d72077cf96f31ed1f10abd7fc23c016d8fbd1c65c73a6128d44fc5f64c9f765" Dec 17 08:30:46 crc kubenswrapper[4966]: I1217 08:30:46.808356 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:30:46 crc kubenswrapper[4966]: I1217 08:30:46.809041 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:31:16 crc kubenswrapper[4966]: I1217 08:31:16.807646 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:31:16 crc kubenswrapper[4966]: I1217 08:31:16.808176 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.167641 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vv5nt"] Dec 17 08:31:37 crc kubenswrapper[4966]: E1217 08:31:37.168323 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2" containerName="collect-profiles" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.168353 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2" containerName="collect-profiles" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.168448 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2" containerName="collect-profiles" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.168841 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-vv5nt" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.174138 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.174999 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.176406 4966 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-jvxfk" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.178439 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vv5nt"] Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.184852 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-zs5rd"] Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.185533 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-zs5rd" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.187920 4966 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-8z24k" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.197509 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-zs5rd"] Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.215031 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-gsbct"] Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.215754 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.220286 4966 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-6j5cj" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.246285 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-gsbct"] Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.302541 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgbd7\" (UniqueName: \"kubernetes.io/projected/a043da8b-6da4-43c6-8015-1342248a8654-kube-api-access-wgbd7\") pod \"cert-manager-5b446d88c5-zs5rd\" (UID: \"a043da8b-6da4-43c6-8015-1342248a8654\") " pod="cert-manager/cert-manager-5b446d88c5-zs5rd" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.302591 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n9cg\" (UniqueName: \"kubernetes.io/projected/959cb714-4a20-4f2b-9145-2d76b29bf94f-kube-api-access-7n9cg\") pod \"cert-manager-cainjector-7f985d654d-vv5nt\" (UID: \"959cb714-4a20-4f2b-9145-2d76b29bf94f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vv5nt" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.403744 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n9cg\" (UniqueName: \"kubernetes.io/projected/959cb714-4a20-4f2b-9145-2d76b29bf94f-kube-api-access-7n9cg\") pod \"cert-manager-cainjector-7f985d654d-vv5nt\" (UID: \"959cb714-4a20-4f2b-9145-2d76b29bf94f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vv5nt" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.403819 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf52t\" (UniqueName: \"kubernetes.io/projected/ad1db831-7ead-45f8-914a-bc49655e4a36-kube-api-access-hf52t\") pod \"cert-manager-webhook-5655c58dd6-gsbct\" (UID: \"ad1db831-7ead-45f8-914a-bc49655e4a36\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.403958 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgbd7\" (UniqueName: \"kubernetes.io/projected/a043da8b-6da4-43c6-8015-1342248a8654-kube-api-access-wgbd7\") pod \"cert-manager-5b446d88c5-zs5rd\" (UID: \"a043da8b-6da4-43c6-8015-1342248a8654\") " pod="cert-manager/cert-manager-5b446d88c5-zs5rd" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.429756 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n9cg\" (UniqueName: \"kubernetes.io/projected/959cb714-4a20-4f2b-9145-2d76b29bf94f-kube-api-access-7n9cg\") pod \"cert-manager-cainjector-7f985d654d-vv5nt\" (UID: \"959cb714-4a20-4f2b-9145-2d76b29bf94f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-vv5nt" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.430644 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgbd7\" (UniqueName: \"kubernetes.io/projected/a043da8b-6da4-43c6-8015-1342248a8654-kube-api-access-wgbd7\") pod \"cert-manager-5b446d88c5-zs5rd\" (UID: \"a043da8b-6da4-43c6-8015-1342248a8654\") " pod="cert-manager/cert-manager-5b446d88c5-zs5rd" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.487582 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-vv5nt" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.506658 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf52t\" (UniqueName: \"kubernetes.io/projected/ad1db831-7ead-45f8-914a-bc49655e4a36-kube-api-access-hf52t\") pod \"cert-manager-webhook-5655c58dd6-gsbct\" (UID: \"ad1db831-7ead-45f8-914a-bc49655e4a36\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.507191 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-zs5rd" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.522439 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf52t\" (UniqueName: \"kubernetes.io/projected/ad1db831-7ead-45f8-914a-bc49655e4a36-kube-api-access-hf52t\") pod \"cert-manager-webhook-5655c58dd6-gsbct\" (UID: \"ad1db831-7ead-45f8-914a-bc49655e4a36\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.530273 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.691425 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-vv5nt"] Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.698740 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.768097 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-zs5rd"] Dec 17 08:31:37 crc kubenswrapper[4966]: I1217 08:31:37.811846 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-gsbct"] Dec 17 08:31:37 crc kubenswrapper[4966]: W1217 08:31:37.815749 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad1db831_7ead_45f8_914a_bc49655e4a36.slice/crio-735e6db07284d6efcde20011e561d9cd681e6f3885be37a61b48c6385ec4f15f WatchSource:0}: Error finding container 735e6db07284d6efcde20011e561d9cd681e6f3885be37a61b48c6385ec4f15f: Status 404 returned error can't find the container with id 735e6db07284d6efcde20011e561d9cd681e6f3885be37a61b48c6385ec4f15f Dec 17 08:31:38 crc kubenswrapper[4966]: I1217 08:31:38.115664 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-vv5nt" event={"ID":"959cb714-4a20-4f2b-9145-2d76b29bf94f","Type":"ContainerStarted","Data":"463d4ff7720d379fb0314c384af5bb466a539d7af3aaab9c2c39f90a42db2518"} Dec 17 08:31:38 crc kubenswrapper[4966]: I1217 08:31:38.116417 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" event={"ID":"ad1db831-7ead-45f8-914a-bc49655e4a36","Type":"ContainerStarted","Data":"735e6db07284d6efcde20011e561d9cd681e6f3885be37a61b48c6385ec4f15f"} Dec 17 08:31:38 crc kubenswrapper[4966]: I1217 08:31:38.117285 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-zs5rd" event={"ID":"a043da8b-6da4-43c6-8015-1342248a8654","Type":"ContainerStarted","Data":"e138019b77113fb6c3379d4ec6a785beae6b00de9cf27e15b3dae4a6596e2347"} Dec 17 08:31:40 crc kubenswrapper[4966]: I1217 08:31:40.224406 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-vv5nt" event={"ID":"959cb714-4a20-4f2b-9145-2d76b29bf94f","Type":"ContainerStarted","Data":"d8888b17639623fca41c41b0743e3b8a716ca374c20b9d0e7a781098dff42975"} Dec 17 08:31:40 crc kubenswrapper[4966]: I1217 08:31:40.256988 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-vv5nt" podStartSLOduration=1.309152524 podStartE2EDuration="3.256964207s" podCreationTimestamp="2025-12-17 08:31:37 +0000 UTC" firstStartedPulling="2025-12-17 08:31:37.698531253 +0000 UTC m=+633.243601195" lastFinishedPulling="2025-12-17 08:31:39.646342936 +0000 UTC m=+635.191412878" observedRunningTime="2025-12-17 08:31:40.24620727 +0000 UTC m=+635.791277212" watchObservedRunningTime="2025-12-17 08:31:40.256964207 +0000 UTC m=+635.802034149" Dec 17 08:31:42 crc kubenswrapper[4966]: I1217 08:31:42.248320 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" event={"ID":"ad1db831-7ead-45f8-914a-bc49655e4a36","Type":"ContainerStarted","Data":"6ff93f31129a753cf2036a3ce7238e76780bba6f5e5cf178a1053c2831ec620d"} Dec 17 08:31:42 crc kubenswrapper[4966]: I1217 08:31:42.248747 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" Dec 17 08:31:42 crc kubenswrapper[4966]: I1217 08:31:42.253422 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-zs5rd" event={"ID":"a043da8b-6da4-43c6-8015-1342248a8654","Type":"ContainerStarted","Data":"b77323ecb0f955e2c31a8f683d388963023faa0037c3221783da57a4d942913e"} Dec 17 08:31:42 crc kubenswrapper[4966]: I1217 08:31:42.268578 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" podStartSLOduration=1.640831683 podStartE2EDuration="5.268559059s" podCreationTimestamp="2025-12-17 08:31:37 +0000 UTC" firstStartedPulling="2025-12-17 08:31:37.818687007 +0000 UTC m=+633.363756949" lastFinishedPulling="2025-12-17 08:31:41.446414383 +0000 UTC m=+636.991484325" observedRunningTime="2025-12-17 08:31:42.267393687 +0000 UTC m=+637.812463629" watchObservedRunningTime="2025-12-17 08:31:42.268559059 +0000 UTC m=+637.813629001" Dec 17 08:31:42 crc kubenswrapper[4966]: I1217 08:31:42.300813 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-zs5rd" podStartSLOduration=1.647019053 podStartE2EDuration="5.300782238s" podCreationTimestamp="2025-12-17 08:31:37 +0000 UTC" firstStartedPulling="2025-12-17 08:31:37.786297863 +0000 UTC m=+633.331367815" lastFinishedPulling="2025-12-17 08:31:41.440061058 +0000 UTC m=+636.985131000" observedRunningTime="2025-12-17 08:31:42.298265068 +0000 UTC m=+637.843335020" watchObservedRunningTime="2025-12-17 08:31:42.300782238 +0000 UTC m=+637.845852220" Dec 17 08:31:46 crc kubenswrapper[4966]: I1217 08:31:46.807962 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:31:46 crc kubenswrapper[4966]: I1217 08:31:46.808461 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:31:46 crc kubenswrapper[4966]: I1217 08:31:46.808507 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:31:46 crc kubenswrapper[4966]: I1217 08:31:46.809157 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee72e03cd88492b874cbcd116164fd7756b1677dac83cef997b128cc487a5628"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 08:31:46 crc kubenswrapper[4966]: I1217 08:31:46.809229 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://ee72e03cd88492b874cbcd116164fd7756b1677dac83cef997b128cc487a5628" gracePeriod=600 Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.285952 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="ee72e03cd88492b874cbcd116164fd7756b1677dac83cef997b128cc487a5628" exitCode=0 Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.286300 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"ee72e03cd88492b874cbcd116164fd7756b1677dac83cef997b128cc487a5628"} Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.286328 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"6002081a3e53c22c985a84be589971720349cfd91c47bfefcfa37db148bb69f7"} Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.286344 4966 scope.go:117] "RemoveContainer" containerID="af9614c017f65ecd41597b9880f18912c6650c9ea6e2d776d690cac78b26db56" Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.490236 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6lq7n"] Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.490676 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovn-controller" containerID="cri-o://1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a" gracePeriod=30 Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.491093 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="sbdb" containerID="cri-o://8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f" gracePeriod=30 Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.491154 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="nbdb" containerID="cri-o://f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726" gracePeriod=30 Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.491206 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="northd" containerID="cri-o://9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384" gracePeriod=30 Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.491248 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd" gracePeriod=30 Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.491291 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kube-rbac-proxy-node" containerID="cri-o://44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add" gracePeriod=30 Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.491331 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovn-acl-logging" containerID="cri-o://ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4" gracePeriod=30 Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.541276 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-gsbct" Dec 17 08:31:47 crc kubenswrapper[4966]: I1217 08:31:47.545471 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" containerID="cri-o://91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c" gracePeriod=30 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.236360 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/3.log" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.238723 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovn-acl-logging/0.log" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.239298 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovn-controller/0.log" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.239747 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.287720 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cnj8n"] Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.287962 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="nbdb" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.287978 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="nbdb" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.287988 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.287994 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288002 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="northd" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288009 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="northd" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288019 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kubecfg-setup" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288025 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kubecfg-setup" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288034 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288039 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288046 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288052 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288061 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288067 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288073 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kube-rbac-proxy-ovn-metrics" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288078 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kube-rbac-proxy-ovn-metrics" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288086 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovn-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288091 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovn-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288099 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovn-acl-logging" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288105 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovn-acl-logging" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288117 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kube-rbac-proxy-node" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288124 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kube-rbac-proxy-node" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288135 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="sbdb" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288141 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="sbdb" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288252 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovn-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288264 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288273 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288280 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kube-rbac-proxy-node" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288287 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="nbdb" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288293 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="sbdb" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288300 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="kube-rbac-proxy-ovn-metrics" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288309 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovn-acl-logging" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288316 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288324 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="northd" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.288442 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288449 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288540 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.288551 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerName="ovnkube-controller" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.290755 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.297179 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/2.log" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.298084 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/1.log" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.298138 4966 generic.go:334] "Generic (PLEG): container finished" podID="ac204c30-92c9-4b1d-9cda-bef16b7a0008" containerID="66f24290075e5dc241bbf1d65160b46bebcbca3bbfb1204419f82204f5c2dde4" exitCode=2 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.298205 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqk4j" event={"ID":"ac204c30-92c9-4b1d-9cda-bef16b7a0008","Type":"ContainerDied","Data":"66f24290075e5dc241bbf1d65160b46bebcbca3bbfb1204419f82204f5c2dde4"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.298250 4966 scope.go:117] "RemoveContainer" containerID="2649479606c9b6e4ad9c7fd7c4085b1ce4381321e6ef13e39452663625adc4c1" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.298736 4966 scope.go:117] "RemoveContainer" containerID="66f24290075e5dc241bbf1d65160b46bebcbca3bbfb1204419f82204f5c2dde4" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.299058 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-lqk4j_openshift-multus(ac204c30-92c9-4b1d-9cda-bef16b7a0008)\"" pod="openshift-multus/multus-lqk4j" podUID="ac204c30-92c9-4b1d-9cda-bef16b7a0008" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.314563 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovnkube-controller/3.log" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.316712 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovn-acl-logging/0.log" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317347 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6lq7n_865dfc88-97a7-4cdb-9564-f46cfadae5dd/ovn-controller/0.log" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317750 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c" exitCode=0 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317780 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f" exitCode=0 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317788 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726" exitCode=0 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317798 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384" exitCode=0 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317805 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd" exitCode=0 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317812 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add" exitCode=0 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317819 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4" exitCode=143 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317826 4966 generic.go:334] "Generic (PLEG): container finished" podID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" containerID="1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a" exitCode=143 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317848 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317887 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317900 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317912 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317920 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317923 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.317930 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318473 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318493 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318503 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318511 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318520 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318527 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318534 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318544 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318551 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318558 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318575 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318590 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318599 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318607 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318614 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318621 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318629 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318635 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318643 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318650 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318657 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318666 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318676 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318684 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318691 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318698 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318704 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318712 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318719 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318726 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318733 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318739 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318748 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6lq7n" event={"ID":"865dfc88-97a7-4cdb-9564-f46cfadae5dd","Type":"ContainerDied","Data":"6ea787bbe772e55ac5e4a7d96f33a23f2e2ed2ddc1f3991cc392b93dd87008c0"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318760 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318768 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318775 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318782 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318789 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318795 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318803 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318822 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318828 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.318835 4966 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483"} Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.339052 4966 scope.go:117] "RemoveContainer" containerID="91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343188 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-var-lib-openvswitch\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343227 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-log-socket\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343256 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-script-lib\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343270 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-netd\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343286 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-ovn-kubernetes\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343304 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-systemd\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343329 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-ovn\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343351 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-openvswitch\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343367 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-netns\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343379 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-kubelet\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343396 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-env-overrides\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343417 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovn-node-metrics-cert\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343434 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-node-log\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343453 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-slash\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343466 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-bin\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343496 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-config\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343516 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-etc-openvswitch\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343538 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-systemd-units\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343564 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnvtq\" (UniqueName: \"kubernetes.io/projected/865dfc88-97a7-4cdb-9564-f46cfadae5dd-kube-api-access-vnvtq\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343584 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\" (UID: \"865dfc88-97a7-4cdb-9564-f46cfadae5dd\") " Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343785 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343812 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.343830 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-log-socket" (OuterVolumeSpecName: "log-socket") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344139 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344150 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344177 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344366 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344412 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344441 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344457 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344496 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344524 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344547 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344569 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344591 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-node-log" (OuterVolumeSpecName: "node-log") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.344610 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-slash" (OuterVolumeSpecName: "host-slash") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.345103 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.350258 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.350339 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/865dfc88-97a7-4cdb-9564-f46cfadae5dd-kube-api-access-vnvtq" (OuterVolumeSpecName: "kube-api-access-vnvtq") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "kube-api-access-vnvtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.362762 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "865dfc88-97a7-4cdb-9564-f46cfadae5dd" (UID: "865dfc88-97a7-4cdb-9564-f46cfadae5dd"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.390306 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.405200 4966 scope.go:117] "RemoveContainer" containerID="8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.416523 4966 scope.go:117] "RemoveContainer" containerID="f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444600 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-cni-bin\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444642 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcvqd\" (UniqueName: \"kubernetes.io/projected/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-kube-api-access-qcvqd\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444666 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444722 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-ovnkube-script-lib\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444785 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-log-socket\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444843 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-systemd-units\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444864 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-slash\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444918 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-env-overrides\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444934 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-run-systemd\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444946 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-run-netns\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444964 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-run-ovn\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.444981 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-ovnkube-config\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445004 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-etc-openvswitch\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445033 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-run-ovn-kubernetes\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445046 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-cni-netd\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445060 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-var-lib-openvswitch\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445084 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-node-log\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445098 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-run-openvswitch\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445112 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-ovn-node-metrics-cert\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445128 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-kubelet\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445164 4966 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445175 4966 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445184 4966 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-log-socket\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445193 4966 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445202 4966 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445210 4966 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445246 4966 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445255 4966 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445264 4966 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445272 4966 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445280 4966 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445288 4966 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445296 4966 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445323 4966 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-node-log\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445335 4966 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-slash\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445343 4966 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445351 4966 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/865dfc88-97a7-4cdb-9564-f46cfadae5dd-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445359 4966 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445367 4966 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/865dfc88-97a7-4cdb-9564-f46cfadae5dd-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.445375 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnvtq\" (UniqueName: \"kubernetes.io/projected/865dfc88-97a7-4cdb-9564-f46cfadae5dd-kube-api-access-vnvtq\") on node \"crc\" DevicePath \"\"" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.448077 4966 scope.go:117] "RemoveContainer" containerID="9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.461770 4966 scope.go:117] "RemoveContainer" containerID="777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.479431 4966 scope.go:117] "RemoveContainer" containerID="44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.494197 4966 scope.go:117] "RemoveContainer" containerID="ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.507282 4966 scope.go:117] "RemoveContainer" containerID="1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.521278 4966 scope.go:117] "RemoveContainer" containerID="cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.537546 4966 scope.go:117] "RemoveContainer" containerID="91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.538664 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c\": container with ID starting with 91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c not found: ID does not exist" containerID="91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.538696 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} err="failed to get container status \"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c\": rpc error: code = NotFound desc = could not find container \"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c\": container with ID starting with 91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.538718 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.539233 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\": container with ID starting with ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f not found: ID does not exist" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.539260 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} err="failed to get container status \"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\": rpc error: code = NotFound desc = could not find container \"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\": container with ID starting with ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.539280 4966 scope.go:117] "RemoveContainer" containerID="8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.539777 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\": container with ID starting with 8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f not found: ID does not exist" containerID="8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.539826 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} err="failed to get container status \"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\": rpc error: code = NotFound desc = could not find container \"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\": container with ID starting with 8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.539859 4966 scope.go:117] "RemoveContainer" containerID="f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.540209 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\": container with ID starting with f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726 not found: ID does not exist" containerID="f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.540234 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} err="failed to get container status \"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\": rpc error: code = NotFound desc = could not find container \"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\": container with ID starting with f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.540253 4966 scope.go:117] "RemoveContainer" containerID="9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.540502 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\": container with ID starting with 9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384 not found: ID does not exist" containerID="9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.540529 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} err="failed to get container status \"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\": rpc error: code = NotFound desc = could not find container \"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\": container with ID starting with 9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.540543 4966 scope.go:117] "RemoveContainer" containerID="777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.540882 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\": container with ID starting with 777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd not found: ID does not exist" containerID="777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.540903 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} err="failed to get container status \"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\": rpc error: code = NotFound desc = could not find container \"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\": container with ID starting with 777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.540914 4966 scope.go:117] "RemoveContainer" containerID="44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.541250 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\": container with ID starting with 44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add not found: ID does not exist" containerID="44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.541299 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} err="failed to get container status \"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\": rpc error: code = NotFound desc = could not find container \"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\": container with ID starting with 44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.541315 4966 scope.go:117] "RemoveContainer" containerID="ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.541581 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\": container with ID starting with ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4 not found: ID does not exist" containerID="ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.541600 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} err="failed to get container status \"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\": rpc error: code = NotFound desc = could not find container \"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\": container with ID starting with ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.541611 4966 scope.go:117] "RemoveContainer" containerID="1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.541897 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\": container with ID starting with 1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a not found: ID does not exist" containerID="1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.541933 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} err="failed to get container status \"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\": rpc error: code = NotFound desc = could not find container \"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\": container with ID starting with 1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.541952 4966 scope.go:117] "RemoveContainer" containerID="cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483" Dec 17 08:31:48 crc kubenswrapper[4966]: E1217 08:31:48.542229 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\": container with ID starting with cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483 not found: ID does not exist" containerID="cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.542253 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483"} err="failed to get container status \"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\": rpc error: code = NotFound desc = could not find container \"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\": container with ID starting with cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.542266 4966 scope.go:117] "RemoveContainer" containerID="91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.542435 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} err="failed to get container status \"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c\": rpc error: code = NotFound desc = could not find container \"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c\": container with ID starting with 91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.542452 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.542602 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} err="failed to get container status \"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\": rpc error: code = NotFound desc = could not find container \"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\": container with ID starting with ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.542620 4966 scope.go:117] "RemoveContainer" containerID="8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.543404 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} err="failed to get container status \"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\": rpc error: code = NotFound desc = could not find container \"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\": container with ID starting with 8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.543425 4966 scope.go:117] "RemoveContainer" containerID="f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.543609 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} err="failed to get container status \"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\": rpc error: code = NotFound desc = could not find container \"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\": container with ID starting with f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.543633 4966 scope.go:117] "RemoveContainer" containerID="9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.543810 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} err="failed to get container status \"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\": rpc error: code = NotFound desc = could not find container \"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\": container with ID starting with 9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.543832 4966 scope.go:117] "RemoveContainer" containerID="777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.544041 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} err="failed to get container status \"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\": rpc error: code = NotFound desc = could not find container \"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\": container with ID starting with 777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.544064 4966 scope.go:117] "RemoveContainer" containerID="44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.544234 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} err="failed to get container status \"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\": rpc error: code = NotFound desc = could not find container \"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\": container with ID starting with 44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.544253 4966 scope.go:117] "RemoveContainer" containerID="ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.544466 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} err="failed to get container status \"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\": rpc error: code = NotFound desc = could not find container \"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\": container with ID starting with ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.544508 4966 scope.go:117] "RemoveContainer" containerID="1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.544713 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} err="failed to get container status \"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\": rpc error: code = NotFound desc = could not find container \"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\": container with ID starting with 1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.544736 4966 scope.go:117] "RemoveContainer" containerID="cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.544984 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483"} err="failed to get container status \"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\": rpc error: code = NotFound desc = could not find container \"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\": container with ID starting with cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545001 4966 scope.go:117] "RemoveContainer" containerID="91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545195 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} err="failed to get container status \"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c\": rpc error: code = NotFound desc = could not find container \"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c\": container with ID starting with 91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545228 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545465 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} err="failed to get container status \"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\": rpc error: code = NotFound desc = could not find container \"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\": container with ID starting with ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545489 4966 scope.go:117] "RemoveContainer" containerID="8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545691 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} err="failed to get container status \"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\": rpc error: code = NotFound desc = could not find container \"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\": container with ID starting with 8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545714 4966 scope.go:117] "RemoveContainer" containerID="f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545902 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} err="failed to get container status \"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\": rpc error: code = NotFound desc = could not find container \"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\": container with ID starting with f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545914 4966 scope.go:117] "RemoveContainer" containerID="9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545940 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-node-log\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545961 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-ovn-node-metrics-cert\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545977 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-run-openvswitch\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.545992 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-kubelet\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546010 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-cni-bin\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546023 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-node-log\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546029 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-run-openvswitch\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546031 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcvqd\" (UniqueName: \"kubernetes.io/projected/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-kube-api-access-qcvqd\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546084 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} err="failed to get container status \"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\": rpc error: code = NotFound desc = could not find container \"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\": container with ID starting with 9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546097 4966 scope.go:117] "RemoveContainer" containerID="777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546099 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546126 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-ovnkube-script-lib\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546150 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546186 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-log-socket\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546269 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-systemd-units\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546306 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-slash\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546403 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-env-overrides\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546430 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-run-systemd\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546457 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-run-netns\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546479 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-run-ovn\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546501 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-ovnkube-config\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546533 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-etc-openvswitch\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546584 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-run-ovn-kubernetes\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546604 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-cni-netd\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546625 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-var-lib-openvswitch\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546726 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-ovnkube-script-lib\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546730 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-var-lib-openvswitch\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546761 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-log-socket\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546781 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} err="failed to get container status \"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\": rpc error: code = NotFound desc = could not find container \"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\": container with ID starting with 777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546789 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-systemd-units\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546795 4966 scope.go:117] "RemoveContainer" containerID="44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546827 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-kubelet\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546850 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-run-ovn\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.546889 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-slash\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547430 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-env-overrides\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547469 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-run-systemd\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547491 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-run-netns\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547535 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-run-ovn-kubernetes\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547559 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-etc-openvswitch\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547582 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-cni-bin\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547601 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-host-cni-netd\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547628 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-ovnkube-config\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547647 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} err="failed to get container status \"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\": rpc error: code = NotFound desc = could not find container \"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\": container with ID starting with 44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.547661 4966 scope.go:117] "RemoveContainer" containerID="ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.548423 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} err="failed to get container status \"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\": rpc error: code = NotFound desc = could not find container \"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\": container with ID starting with ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.548437 4966 scope.go:117] "RemoveContainer" containerID="1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549011 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} err="failed to get container status \"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\": rpc error: code = NotFound desc = could not find container \"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\": container with ID starting with 1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549033 4966 scope.go:117] "RemoveContainer" containerID="cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549230 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483"} err="failed to get container status \"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\": rpc error: code = NotFound desc = could not find container \"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\": container with ID starting with cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549247 4966 scope.go:117] "RemoveContainer" containerID="91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549317 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-ovn-node-metrics-cert\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549434 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c"} err="failed to get container status \"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c\": rpc error: code = NotFound desc = could not find container \"91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c\": container with ID starting with 91518487ffb796bd4c0118163c3ac7edf83b79909da5ece131515e3f395fdb4c not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549453 4966 scope.go:117] "RemoveContainer" containerID="ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549605 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f"} err="failed to get container status \"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\": rpc error: code = NotFound desc = could not find container \"ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f\": container with ID starting with ffc95ea1f9895a91760186a9f8b111c92d136734425b75bd436bc430a4f5454f not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549621 4966 scope.go:117] "RemoveContainer" containerID="8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549768 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f"} err="failed to get container status \"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\": rpc error: code = NotFound desc = could not find container \"8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f\": container with ID starting with 8af3f845b4fe7a809eaf27d303194cf56a977677fa8fa7c127de62de230e0c6f not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.549784 4966 scope.go:117] "RemoveContainer" containerID="f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550020 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726"} err="failed to get container status \"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\": rpc error: code = NotFound desc = could not find container \"f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726\": container with ID starting with f2e1b3a62ae3c49cbc7095dacf1ab6d91d2cb9c8a757dcbd9245f6ea523da726 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550038 4966 scope.go:117] "RemoveContainer" containerID="9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550214 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384"} err="failed to get container status \"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\": rpc error: code = NotFound desc = could not find container \"9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384\": container with ID starting with 9d69656d42aeaa0ce930764436e8232c340f8b11229f134b30bc054e09083384 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550231 4966 scope.go:117] "RemoveContainer" containerID="777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550385 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd"} err="failed to get container status \"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\": rpc error: code = NotFound desc = could not find container \"777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd\": container with ID starting with 777aa7d9e246e163eeebe4b97d9b24c9a9b0bed3bbfe91b6cb634117df51f7fd not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550404 4966 scope.go:117] "RemoveContainer" containerID="44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550574 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add"} err="failed to get container status \"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\": rpc error: code = NotFound desc = could not find container \"44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add\": container with ID starting with 44976dc500b7b679328f1f99008790a7910e3ba01b0555378d00b04057aa2add not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550590 4966 scope.go:117] "RemoveContainer" containerID="ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550767 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4"} err="failed to get container status \"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\": rpc error: code = NotFound desc = could not find container \"ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4\": container with ID starting with ee76b88f3dfc5c05ad50990ae1ef6fdd5ee5fa04de91e61466254cc0d85d5bf4 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550784 4966 scope.go:117] "RemoveContainer" containerID="1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.550989 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a"} err="failed to get container status \"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\": rpc error: code = NotFound desc = could not find container \"1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a\": container with ID starting with 1e15195f6d35e233a49b7d62b302940666a8e5561ae390865e7065f430a5f51a not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.551012 4966 scope.go:117] "RemoveContainer" containerID="cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.551207 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483"} err="failed to get container status \"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\": rpc error: code = NotFound desc = could not find container \"cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483\": container with ID starting with cf77807bd4fe5a5138c60fe07c400578f56110ed15ed3bfb1e7612320757f483 not found: ID does not exist" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.560022 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcvqd\" (UniqueName: \"kubernetes.io/projected/5a23ba16-1bcd-4c00-a4f4-30b3a14eab54-kube-api-access-qcvqd\") pod \"ovnkube-node-cnj8n\" (UID: \"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54\") " pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.633375 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.660382 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6lq7n"] Dec 17 08:31:48 crc kubenswrapper[4966]: W1217 08:31:48.664475 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a23ba16_1bcd_4c00_a4f4_30b3a14eab54.slice/crio-4acf06768df25387f3e6ccfd84f52316811ae166e27829c7e63b1c2581c91f38 WatchSource:0}: Error finding container 4acf06768df25387f3e6ccfd84f52316811ae166e27829c7e63b1c2581c91f38: Status 404 returned error can't find the container with id 4acf06768df25387f3e6ccfd84f52316811ae166e27829c7e63b1c2581c91f38 Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.666172 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6lq7n"] Dec 17 08:31:48 crc kubenswrapper[4966]: I1217 08:31:48.846235 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="865dfc88-97a7-4cdb-9564-f46cfadae5dd" path="/var/lib/kubelet/pods/865dfc88-97a7-4cdb-9564-f46cfadae5dd/volumes" Dec 17 08:31:49 crc kubenswrapper[4966]: I1217 08:31:49.326111 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/2.log" Dec 17 08:31:49 crc kubenswrapper[4966]: I1217 08:31:49.328203 4966 generic.go:334] "Generic (PLEG): container finished" podID="5a23ba16-1bcd-4c00-a4f4-30b3a14eab54" containerID="b5ff62967e418c42bcb22b9ae93bd8a27545c93f6097e53df1c45afce9a3df62" exitCode=0 Dec 17 08:31:49 crc kubenswrapper[4966]: I1217 08:31:49.328296 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerDied","Data":"b5ff62967e418c42bcb22b9ae93bd8a27545c93f6097e53df1c45afce9a3df62"} Dec 17 08:31:49 crc kubenswrapper[4966]: I1217 08:31:49.328469 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerStarted","Data":"4acf06768df25387f3e6ccfd84f52316811ae166e27829c7e63b1c2581c91f38"} Dec 17 08:31:50 crc kubenswrapper[4966]: I1217 08:31:50.340980 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerStarted","Data":"1c092dc4e80e4c3c55d593267b56487cf422a6d8e7271150cc1d8161285eba0c"} Dec 17 08:31:50 crc kubenswrapper[4966]: I1217 08:31:50.341061 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerStarted","Data":"edd7bcb09a2f98bb39a44b48d251dfd76a6a917fadce439ad693cb83d8207d72"} Dec 17 08:31:50 crc kubenswrapper[4966]: I1217 08:31:50.341087 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerStarted","Data":"4297bd3672383d20e6a902882ccf5fae180862879a8864ef6cdf8ec920bbfb15"} Dec 17 08:31:50 crc kubenswrapper[4966]: I1217 08:31:50.341127 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerStarted","Data":"1d06a7b0dc0d74e9d6bbe21265248540810aabd9fa89975f235dab12facc4377"} Dec 17 08:31:50 crc kubenswrapper[4966]: I1217 08:31:50.341146 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerStarted","Data":"30b7ce02942882da91c32de33334beac55ff07f617ff3cf2dabda53a46c44616"} Dec 17 08:31:50 crc kubenswrapper[4966]: I1217 08:31:50.341164 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerStarted","Data":"906baee927db92ae81b37ae208118829e533480200271803378ce3165a49328c"} Dec 17 08:31:52 crc kubenswrapper[4966]: I1217 08:31:52.359252 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerStarted","Data":"d03ae3c4ebef8e1163188a012c62c184c20b4f120b2776f06edb93d0170a17e1"} Dec 17 08:31:55 crc kubenswrapper[4966]: I1217 08:31:55.381137 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" event={"ID":"5a23ba16-1bcd-4c00-a4f4-30b3a14eab54","Type":"ContainerStarted","Data":"3d406220b6337cd9e23999262f93e0439056c4c9072922c146a2479977737236"} Dec 17 08:31:55 crc kubenswrapper[4966]: I1217 08:31:55.382097 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:55 crc kubenswrapper[4966]: I1217 08:31:55.382117 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:55 crc kubenswrapper[4966]: I1217 08:31:55.412678 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:55 crc kubenswrapper[4966]: I1217 08:31:55.415991 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" podStartSLOduration=7.415969579 podStartE2EDuration="7.415969579s" podCreationTimestamp="2025-12-17 08:31:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:31:55.411971789 +0000 UTC m=+650.957041751" watchObservedRunningTime="2025-12-17 08:31:55.415969579 +0000 UTC m=+650.961039521" Dec 17 08:31:56 crc kubenswrapper[4966]: I1217 08:31:56.389181 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:31:56 crc kubenswrapper[4966]: I1217 08:31:56.423947 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:32:01 crc kubenswrapper[4966]: I1217 08:32:01.830427 4966 scope.go:117] "RemoveContainer" containerID="66f24290075e5dc241bbf1d65160b46bebcbca3bbfb1204419f82204f5c2dde4" Dec 17 08:32:01 crc kubenswrapper[4966]: E1217 08:32:01.830976 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-lqk4j_openshift-multus(ac204c30-92c9-4b1d-9cda-bef16b7a0008)\"" pod="openshift-multus/multus-lqk4j" podUID="ac204c30-92c9-4b1d-9cda-bef16b7a0008" Dec 17 08:32:14 crc kubenswrapper[4966]: I1217 08:32:14.837987 4966 scope.go:117] "RemoveContainer" containerID="66f24290075e5dc241bbf1d65160b46bebcbca3bbfb1204419f82204f5c2dde4" Dec 17 08:32:15 crc kubenswrapper[4966]: I1217 08:32:15.535409 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqk4j_ac204c30-92c9-4b1d-9cda-bef16b7a0008/kube-multus/2.log" Dec 17 08:32:15 crc kubenswrapper[4966]: I1217 08:32:15.535823 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqk4j" event={"ID":"ac204c30-92c9-4b1d-9cda-bef16b7a0008","Type":"ContainerStarted","Data":"6f1628bf2f0438d25aa80804487c2482759df24671447044dcfc6987b24a9012"} Dec 17 08:32:18 crc kubenswrapper[4966]: I1217 08:32:18.660434 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cnj8n" Dec 17 08:32:29 crc kubenswrapper[4966]: I1217 08:32:29.923347 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td"] Dec 17 08:32:29 crc kubenswrapper[4966]: I1217 08:32:29.924950 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:29 crc kubenswrapper[4966]: I1217 08:32:29.928164 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 17 08:32:29 crc kubenswrapper[4966]: I1217 08:32:29.947449 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td"] Dec 17 08:32:29 crc kubenswrapper[4966]: I1217 08:32:29.995060 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:29 crc kubenswrapper[4966]: I1217 08:32:29.995107 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w9b7\" (UniqueName: \"kubernetes.io/projected/04390eb1-6c1d-4681-abbc-7a139dcac943-kube-api-access-4w9b7\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:29 crc kubenswrapper[4966]: I1217 08:32:29.995192 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:30 crc kubenswrapper[4966]: I1217 08:32:30.095672 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:30 crc kubenswrapper[4966]: I1217 08:32:30.095743 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w9b7\" (UniqueName: \"kubernetes.io/projected/04390eb1-6c1d-4681-abbc-7a139dcac943-kube-api-access-4w9b7\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:30 crc kubenswrapper[4966]: I1217 08:32:30.095857 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:30 crc kubenswrapper[4966]: I1217 08:32:30.096583 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:30 crc kubenswrapper[4966]: I1217 08:32:30.096597 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:30 crc kubenswrapper[4966]: I1217 08:32:30.127799 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w9b7\" (UniqueName: \"kubernetes.io/projected/04390eb1-6c1d-4681-abbc-7a139dcac943-kube-api-access-4w9b7\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:30 crc kubenswrapper[4966]: I1217 08:32:30.250599 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:30 crc kubenswrapper[4966]: I1217 08:32:30.654224 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td"] Dec 17 08:32:31 crc kubenswrapper[4966]: I1217 08:32:31.646280 4966 generic.go:334] "Generic (PLEG): container finished" podID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerID="dbc6b767fb784dd72eb46d9c7e1d1f1fe85331a04f47b59e3dad2af877590ac3" exitCode=0 Dec 17 08:32:31 crc kubenswrapper[4966]: I1217 08:32:31.646339 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" event={"ID":"04390eb1-6c1d-4681-abbc-7a139dcac943","Type":"ContainerDied","Data":"dbc6b767fb784dd72eb46d9c7e1d1f1fe85331a04f47b59e3dad2af877590ac3"} Dec 17 08:32:31 crc kubenswrapper[4966]: I1217 08:32:31.648245 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" event={"ID":"04390eb1-6c1d-4681-abbc-7a139dcac943","Type":"ContainerStarted","Data":"7a98991f62da96acdb349c4d189d756e577eb50743629e5beb063a1ec30e2040"} Dec 17 08:32:33 crc kubenswrapper[4966]: I1217 08:32:33.664835 4966 generic.go:334] "Generic (PLEG): container finished" podID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerID="6999fafb37b2c450ed04357a18644cec94911c482d15c573260d1221412f47af" exitCode=0 Dec 17 08:32:33 crc kubenswrapper[4966]: I1217 08:32:33.665245 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" event={"ID":"04390eb1-6c1d-4681-abbc-7a139dcac943","Type":"ContainerDied","Data":"6999fafb37b2c450ed04357a18644cec94911c482d15c573260d1221412f47af"} Dec 17 08:32:34 crc kubenswrapper[4966]: I1217 08:32:34.675179 4966 generic.go:334] "Generic (PLEG): container finished" podID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerID="518de899cfc9d1724438c8322075e3645a9831b649d6649831b102dc3d5e39b9" exitCode=0 Dec 17 08:32:34 crc kubenswrapper[4966]: I1217 08:32:34.675248 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" event={"ID":"04390eb1-6c1d-4681-abbc-7a139dcac943","Type":"ContainerDied","Data":"518de899cfc9d1724438c8322075e3645a9831b649d6649831b102dc3d5e39b9"} Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.088168 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.276831 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-bundle\") pod \"04390eb1-6c1d-4681-abbc-7a139dcac943\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.276912 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w9b7\" (UniqueName: \"kubernetes.io/projected/04390eb1-6c1d-4681-abbc-7a139dcac943-kube-api-access-4w9b7\") pod \"04390eb1-6c1d-4681-abbc-7a139dcac943\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.276967 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-util\") pod \"04390eb1-6c1d-4681-abbc-7a139dcac943\" (UID: \"04390eb1-6c1d-4681-abbc-7a139dcac943\") " Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.277625 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-bundle" (OuterVolumeSpecName: "bundle") pod "04390eb1-6c1d-4681-abbc-7a139dcac943" (UID: "04390eb1-6c1d-4681-abbc-7a139dcac943"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.287032 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04390eb1-6c1d-4681-abbc-7a139dcac943-kube-api-access-4w9b7" (OuterVolumeSpecName: "kube-api-access-4w9b7") pod "04390eb1-6c1d-4681-abbc-7a139dcac943" (UID: "04390eb1-6c1d-4681-abbc-7a139dcac943"). InnerVolumeSpecName "kube-api-access-4w9b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.352232 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-util" (OuterVolumeSpecName: "util") pod "04390eb1-6c1d-4681-abbc-7a139dcac943" (UID: "04390eb1-6c1d-4681-abbc-7a139dcac943"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.378244 4966 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.378295 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w9b7\" (UniqueName: \"kubernetes.io/projected/04390eb1-6c1d-4681-abbc-7a139dcac943-kube-api-access-4w9b7\") on node \"crc\" DevicePath \"\"" Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.378310 4966 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/04390eb1-6c1d-4681-abbc-7a139dcac943-util\") on node \"crc\" DevicePath \"\"" Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.689454 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" event={"ID":"04390eb1-6c1d-4681-abbc-7a139dcac943","Type":"ContainerDied","Data":"7a98991f62da96acdb349c4d189d756e577eb50743629e5beb063a1ec30e2040"} Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.689496 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a98991f62da96acdb349c4d189d756e577eb50743629e5beb063a1ec30e2040" Dec 17 08:32:36 crc kubenswrapper[4966]: I1217 08:32:36.689513 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8mm9td" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.780497 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-gmfsd"] Dec 17 08:32:38 crc kubenswrapper[4966]: E1217 08:32:38.781050 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerName="extract" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.781065 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerName="extract" Dec 17 08:32:38 crc kubenswrapper[4966]: E1217 08:32:38.781090 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerName="pull" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.781098 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerName="pull" Dec 17 08:32:38 crc kubenswrapper[4966]: E1217 08:32:38.781112 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerName="util" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.781119 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerName="util" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.781234 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="04390eb1-6c1d-4681-abbc-7a139dcac943" containerName="extract" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.781732 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-gmfsd" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.797678 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.798336 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.799804 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-47962" Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.857776 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-gmfsd"] Dec 17 08:32:38 crc kubenswrapper[4966]: I1217 08:32:38.912226 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mskrb\" (UniqueName: \"kubernetes.io/projected/0d887962-bd5e-4131-959d-1f8bdbc6ee76-kube-api-access-mskrb\") pod \"nmstate-operator-6769fb99d-gmfsd\" (UID: \"0d887962-bd5e-4131-959d-1f8bdbc6ee76\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-gmfsd" Dec 17 08:32:39 crc kubenswrapper[4966]: I1217 08:32:39.013344 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mskrb\" (UniqueName: \"kubernetes.io/projected/0d887962-bd5e-4131-959d-1f8bdbc6ee76-kube-api-access-mskrb\") pod \"nmstate-operator-6769fb99d-gmfsd\" (UID: \"0d887962-bd5e-4131-959d-1f8bdbc6ee76\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-gmfsd" Dec 17 08:32:39 crc kubenswrapper[4966]: I1217 08:32:39.030171 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mskrb\" (UniqueName: \"kubernetes.io/projected/0d887962-bd5e-4131-959d-1f8bdbc6ee76-kube-api-access-mskrb\") pod \"nmstate-operator-6769fb99d-gmfsd\" (UID: \"0d887962-bd5e-4131-959d-1f8bdbc6ee76\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-gmfsd" Dec 17 08:32:39 crc kubenswrapper[4966]: I1217 08:32:39.115905 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-gmfsd" Dec 17 08:32:39 crc kubenswrapper[4966]: I1217 08:32:39.343055 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-gmfsd"] Dec 17 08:32:39 crc kubenswrapper[4966]: I1217 08:32:39.706265 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-gmfsd" event={"ID":"0d887962-bd5e-4131-959d-1f8bdbc6ee76","Type":"ContainerStarted","Data":"4f2121581e7939bdb16aa198496bcb6f1beb2a6d83bfcd6e2c8eb69ff3ff8411"} Dec 17 08:32:42 crc kubenswrapper[4966]: I1217 08:32:42.723642 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-gmfsd" event={"ID":"0d887962-bd5e-4131-959d-1f8bdbc6ee76","Type":"ContainerStarted","Data":"f754ddf45ff5c70376c293b26939de01c70139ef5d9aabb5fcbd8e19b93b7714"} Dec 17 08:32:42 crc kubenswrapper[4966]: I1217 08:32:42.756959 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-6769fb99d-gmfsd" podStartSLOduration=1.9405764909999998 podStartE2EDuration="4.756931593s" podCreationTimestamp="2025-12-17 08:32:38 +0000 UTC" firstStartedPulling="2025-12-17 08:32:39.350556799 +0000 UTC m=+694.895626741" lastFinishedPulling="2025-12-17 08:32:42.166911901 +0000 UTC m=+697.711981843" observedRunningTime="2025-12-17 08:32:42.738069194 +0000 UTC m=+698.283139176" watchObservedRunningTime="2025-12-17 08:32:42.756931593 +0000 UTC m=+698.302001535" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.640619 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc"] Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.641513 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.643189 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-g4kb6" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.664512 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc"] Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.679665 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k"] Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.680581 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.682159 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.712773 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-j2f2z"] Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.713493 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.735788 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k"] Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.786031 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdqww\" (UniqueName: \"kubernetes.io/projected/895b6a41-d94c-44a2-a12f-47ca636bd482-kube-api-access-mdqww\") pod \"nmstate-webhook-f8fb84555-2wr4k\" (UID: \"895b6a41-d94c-44a2-a12f-47ca636bd482\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.786093 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/895b6a41-d94c-44a2-a12f-47ca636bd482-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-2wr4k\" (UID: \"895b6a41-d94c-44a2-a12f-47ca636bd482\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.786139 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx7xh\" (UniqueName: \"kubernetes.io/projected/53e82fa5-de70-4ee3-85f2-727d9ae3c98a-kube-api-access-zx7xh\") pod \"nmstate-metrics-7f7f7578db-92lfc\" (UID: \"53e82fa5-de70-4ee3-85f2-727d9ae3c98a\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.819119 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj"] Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.820172 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.822177 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-p5xgx" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.822177 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.825763 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.846421 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj"] Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.890462 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/895b6a41-d94c-44a2-a12f-47ca636bd482-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-2wr4k\" (UID: \"895b6a41-d94c-44a2-a12f-47ca636bd482\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.890507 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/de15ebc0-cce8-48be-bd70-97ca8e438cf6-ovs-socket\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.890536 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/de15ebc0-cce8-48be-bd70-97ca8e438cf6-dbus-socket\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.890570 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx7xh\" (UniqueName: \"kubernetes.io/projected/53e82fa5-de70-4ee3-85f2-727d9ae3c98a-kube-api-access-zx7xh\") pod \"nmstate-metrics-7f7f7578db-92lfc\" (UID: \"53e82fa5-de70-4ee3-85f2-727d9ae3c98a\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.890598 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndgth\" (UniqueName: \"kubernetes.io/projected/de15ebc0-cce8-48be-bd70-97ca8e438cf6-kube-api-access-ndgth\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: E1217 08:32:43.890617 4966 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 17 08:32:43 crc kubenswrapper[4966]: E1217 08:32:43.890698 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/895b6a41-d94c-44a2-a12f-47ca636bd482-tls-key-pair podName:895b6a41-d94c-44a2-a12f-47ca636bd482 nodeName:}" failed. No retries permitted until 2025-12-17 08:32:44.390677649 +0000 UTC m=+699.935747681 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/895b6a41-d94c-44a2-a12f-47ca636bd482-tls-key-pair") pod "nmstate-webhook-f8fb84555-2wr4k" (UID: "895b6a41-d94c-44a2-a12f-47ca636bd482") : secret "openshift-nmstate-webhook" not found Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.890639 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/de15ebc0-cce8-48be-bd70-97ca8e438cf6-nmstate-lock\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.891242 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdqww\" (UniqueName: \"kubernetes.io/projected/895b6a41-d94c-44a2-a12f-47ca636bd482-kube-api-access-mdqww\") pod \"nmstate-webhook-f8fb84555-2wr4k\" (UID: \"895b6a41-d94c-44a2-a12f-47ca636bd482\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.910671 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx7xh\" (UniqueName: \"kubernetes.io/projected/53e82fa5-de70-4ee3-85f2-727d9ae3c98a-kube-api-access-zx7xh\") pod \"nmstate-metrics-7f7f7578db-92lfc\" (UID: \"53e82fa5-de70-4ee3-85f2-727d9ae3c98a\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.910767 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdqww\" (UniqueName: \"kubernetes.io/projected/895b6a41-d94c-44a2-a12f-47ca636bd482-kube-api-access-mdqww\") pod \"nmstate-webhook-f8fb84555-2wr4k\" (UID: \"895b6a41-d94c-44a2-a12f-47ca636bd482\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.966359 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.992499 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/de15ebc0-cce8-48be-bd70-97ca8e438cf6-nmstate-lock\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.992579 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/de15ebc0-cce8-48be-bd70-97ca8e438cf6-ovs-socket\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.992627 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6f4076f5-6b3e-43c5-ad15-ed3746c27b3b-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-gndcj\" (UID: \"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.992645 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6f4076f5-6b3e-43c5-ad15-ed3746c27b3b-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-gndcj\" (UID: \"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.992663 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/de15ebc0-cce8-48be-bd70-97ca8e438cf6-dbus-socket\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.992694 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cwxz\" (UniqueName: \"kubernetes.io/projected/6f4076f5-6b3e-43c5-ad15-ed3746c27b3b-kube-api-access-8cwxz\") pod \"nmstate-console-plugin-6ff7998486-gndcj\" (UID: \"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.992717 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndgth\" (UniqueName: \"kubernetes.io/projected/de15ebc0-cce8-48be-bd70-97ca8e438cf6-kube-api-access-ndgth\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.993013 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/de15ebc0-cce8-48be-bd70-97ca8e438cf6-nmstate-lock\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.993032 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/de15ebc0-cce8-48be-bd70-97ca8e438cf6-ovs-socket\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:43 crc kubenswrapper[4966]: I1217 08:32:43.993227 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/de15ebc0-cce8-48be-bd70-97ca8e438cf6-dbus-socket\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.049795 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndgth\" (UniqueName: \"kubernetes.io/projected/de15ebc0-cce8-48be-bd70-97ca8e438cf6-kube-api-access-ndgth\") pod \"nmstate-handler-j2f2z\" (UID: \"de15ebc0-cce8-48be-bd70-97ca8e438cf6\") " pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.091538 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-86cc5994cc-45pt9"] Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.092302 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.093438 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6f4076f5-6b3e-43c5-ad15-ed3746c27b3b-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-gndcj\" (UID: \"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.094021 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6f4076f5-6b3e-43c5-ad15-ed3746c27b3b-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-gndcj\" (UID: \"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.094082 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cwxz\" (UniqueName: \"kubernetes.io/projected/6f4076f5-6b3e-43c5-ad15-ed3746c27b3b-kube-api-access-8cwxz\") pod \"nmstate-console-plugin-6ff7998486-gndcj\" (UID: \"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.095576 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6f4076f5-6b3e-43c5-ad15-ed3746c27b3b-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-gndcj\" (UID: \"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.098907 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6f4076f5-6b3e-43c5-ad15-ed3746c27b3b-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-gndcj\" (UID: \"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.116503 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cwxz\" (UniqueName: \"kubernetes.io/projected/6f4076f5-6b3e-43c5-ad15-ed3746c27b3b-kube-api-access-8cwxz\") pod \"nmstate-console-plugin-6ff7998486-gndcj\" (UID: \"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.135663 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.161282 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86cc5994cc-45pt9"] Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.195092 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-console-config\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.195125 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqw8c\" (UniqueName: \"kubernetes.io/projected/ad2efbe9-6457-427b-8fd9-f7602d339aa3-kube-api-access-pqw8c\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.195149 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-trusted-ca-bundle\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.195179 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-oauth-serving-cert\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.195202 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-service-ca\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.195262 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad2efbe9-6457-427b-8fd9-f7602d339aa3-console-serving-cert\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.195284 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad2efbe9-6457-427b-8fd9-f7602d339aa3-console-oauth-config\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.297005 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-console-config\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.297982 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqw8c\" (UniqueName: \"kubernetes.io/projected/ad2efbe9-6457-427b-8fd9-f7602d339aa3-kube-api-access-pqw8c\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.298007 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-trusted-ca-bundle\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.298023 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-oauth-serving-cert\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.298834 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-service-ca\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.298921 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad2efbe9-6457-427b-8fd9-f7602d339aa3-console-serving-cert\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.298945 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad2efbe9-6457-427b-8fd9-f7602d339aa3-console-oauth-config\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.299194 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-oauth-serving-cert\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.299412 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-trusted-ca-bundle\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.297949 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-console-config\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.300261 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad2efbe9-6457-427b-8fd9-f7602d339aa3-service-ca\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.303245 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad2efbe9-6457-427b-8fd9-f7602d339aa3-console-serving-cert\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.303944 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad2efbe9-6457-427b-8fd9-f7602d339aa3-console-oauth-config\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.316147 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqw8c\" (UniqueName: \"kubernetes.io/projected/ad2efbe9-6457-427b-8fd9-f7602d339aa3-kube-api-access-pqw8c\") pod \"console-86cc5994cc-45pt9\" (UID: \"ad2efbe9-6457-427b-8fd9-f7602d339aa3\") " pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.326060 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.345310 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc"] Dec 17 08:32:44 crc kubenswrapper[4966]: W1217 08:32:44.352076 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53e82fa5_de70_4ee3_85f2_727d9ae3c98a.slice/crio-5a464a6a3c9c41b67759dd7d70b3f4cae9c4ef517cc3205eb8c74fbe87b995be WatchSource:0}: Error finding container 5a464a6a3c9c41b67759dd7d70b3f4cae9c4ef517cc3205eb8c74fbe87b995be: Status 404 returned error can't find the container with id 5a464a6a3c9c41b67759dd7d70b3f4cae9c4ef517cc3205eb8c74fbe87b995be Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.399507 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/895b6a41-d94c-44a2-a12f-47ca636bd482-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-2wr4k\" (UID: \"895b6a41-d94c-44a2-a12f-47ca636bd482\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.402776 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/895b6a41-d94c-44a2-a12f-47ca636bd482-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-2wr4k\" (UID: \"895b6a41-d94c-44a2-a12f-47ca636bd482\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.414642 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.490165 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj"] Dec 17 08:32:44 crc kubenswrapper[4966]: W1217 08:32:44.507179 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f4076f5_6b3e_43c5_ad15_ed3746c27b3b.slice/crio-0eb57845833b3abb8145425d6a16a75331b154f0cd5ab81558ae025e679e59d8 WatchSource:0}: Error finding container 0eb57845833b3abb8145425d6a16a75331b154f0cd5ab81558ae025e679e59d8: Status 404 returned error can't find the container with id 0eb57845833b3abb8145425d6a16a75331b154f0cd5ab81558ae025e679e59d8 Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.597888 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.643954 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86cc5994cc-45pt9"] Dec 17 08:32:44 crc kubenswrapper[4966]: W1217 08:32:44.663816 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad2efbe9_6457_427b_8fd9_f7602d339aa3.slice/crio-036ca427849e5b4082ecc066a36a16d3092699b0362bc11934d1a8583a7da050 WatchSource:0}: Error finding container 036ca427849e5b4082ecc066a36a16d3092699b0362bc11934d1a8583a7da050: Status 404 returned error can't find the container with id 036ca427849e5b4082ecc066a36a16d3092699b0362bc11934d1a8583a7da050 Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.734291 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc" event={"ID":"53e82fa5-de70-4ee3-85f2-727d9ae3c98a","Type":"ContainerStarted","Data":"5a464a6a3c9c41b67759dd7d70b3f4cae9c4ef517cc3205eb8c74fbe87b995be"} Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.737886 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" event={"ID":"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b","Type":"ContainerStarted","Data":"0eb57845833b3abb8145425d6a16a75331b154f0cd5ab81558ae025e679e59d8"} Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.741835 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86cc5994cc-45pt9" event={"ID":"ad2efbe9-6457-427b-8fd9-f7602d339aa3","Type":"ContainerStarted","Data":"036ca427849e5b4082ecc066a36a16d3092699b0362bc11934d1a8583a7da050"} Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.744028 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j2f2z" event={"ID":"de15ebc0-cce8-48be-bd70-97ca8e438cf6","Type":"ContainerStarted","Data":"6c8408be619e0f19ec3fde41543e2b1d31582c8dc994c1f22a966deaeb989d02"} Dec 17 08:32:44 crc kubenswrapper[4966]: I1217 08:32:44.822856 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k"] Dec 17 08:32:45 crc kubenswrapper[4966]: I1217 08:32:45.752252 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" event={"ID":"895b6a41-d94c-44a2-a12f-47ca636bd482","Type":"ContainerStarted","Data":"777a02c29867c712ea39a4507047ccb20f54d2906698caba266307c957eda513"} Dec 17 08:32:45 crc kubenswrapper[4966]: I1217 08:32:45.754309 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86cc5994cc-45pt9" event={"ID":"ad2efbe9-6457-427b-8fd9-f7602d339aa3","Type":"ContainerStarted","Data":"7a018db8ac0a6fb94e3ce0e4aa43b02262734758d3d4c88b3bef32ab4b299f56"} Dec 17 08:32:45 crc kubenswrapper[4966]: I1217 08:32:45.775160 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-86cc5994cc-45pt9" podStartSLOduration=1.775140374 podStartE2EDuration="1.775140374s" podCreationTimestamp="2025-12-17 08:32:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:32:45.772987143 +0000 UTC m=+701.318057085" watchObservedRunningTime="2025-12-17 08:32:45.775140374 +0000 UTC m=+701.320210316" Dec 17 08:32:48 crc kubenswrapper[4966]: I1217 08:32:48.777836 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc" event={"ID":"53e82fa5-de70-4ee3-85f2-727d9ae3c98a","Type":"ContainerStarted","Data":"5c6c4c9ee71eca0e370c6f458f2f78a48eeaddb180d1c35c90a4cfb9ca38d5b3"} Dec 17 08:32:48 crc kubenswrapper[4966]: I1217 08:32:48.779983 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" event={"ID":"6f4076f5-6b3e-43c5-ad15-ed3746c27b3b","Type":"ContainerStarted","Data":"3e5e50deebc50468d4ad46aa358d98b42c25ba1e7565b3623769a94eb4c110e7"} Dec 17 08:32:48 crc kubenswrapper[4966]: I1217 08:32:48.781805 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j2f2z" event={"ID":"de15ebc0-cce8-48be-bd70-97ca8e438cf6","Type":"ContainerStarted","Data":"29f37a5b9dbc556c730c2f91fbc2fac85a127adffbde3c766ef9c73a5e0d7c30"} Dec 17 08:32:48 crc kubenswrapper[4966]: I1217 08:32:48.781908 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:48 crc kubenswrapper[4966]: I1217 08:32:48.783456 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" event={"ID":"895b6a41-d94c-44a2-a12f-47ca636bd482","Type":"ContainerStarted","Data":"31ed7c5d53f0b5ca72243095590ee26b1d06718d306cc91294a5e2039fec37d8"} Dec 17 08:32:48 crc kubenswrapper[4966]: I1217 08:32:48.783602 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:32:48 crc kubenswrapper[4966]: I1217 08:32:48.797802 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-gndcj" podStartSLOduration=1.942365316 podStartE2EDuration="5.797788219s" podCreationTimestamp="2025-12-17 08:32:43 +0000 UTC" firstStartedPulling="2025-12-17 08:32:44.517682879 +0000 UTC m=+700.062752821" lastFinishedPulling="2025-12-17 08:32:48.373105782 +0000 UTC m=+703.918175724" observedRunningTime="2025-12-17 08:32:48.795421013 +0000 UTC m=+704.340490965" watchObservedRunningTime="2025-12-17 08:32:48.797788219 +0000 UTC m=+704.342858161" Dec 17 08:32:48 crc kubenswrapper[4966]: I1217 08:32:48.817786 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" podStartSLOduration=2.278504839 podStartE2EDuration="5.817766079s" podCreationTimestamp="2025-12-17 08:32:43 +0000 UTC" firstStartedPulling="2025-12-17 08:32:44.837736911 +0000 UTC m=+700.382806853" lastFinishedPulling="2025-12-17 08:32:48.376998151 +0000 UTC m=+703.922068093" observedRunningTime="2025-12-17 08:32:48.816890454 +0000 UTC m=+704.361960416" watchObservedRunningTime="2025-12-17 08:32:48.817766079 +0000 UTC m=+704.362836021" Dec 17 08:32:48 crc kubenswrapper[4966]: I1217 08:32:48.839694 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-j2f2z" podStartSLOduration=1.809536402 podStartE2EDuration="5.839676194s" podCreationTimestamp="2025-12-17 08:32:43 +0000 UTC" firstStartedPulling="2025-12-17 08:32:44.364972017 +0000 UTC m=+699.910041959" lastFinishedPulling="2025-12-17 08:32:48.395111779 +0000 UTC m=+703.940181751" observedRunningTime="2025-12-17 08:32:48.839438996 +0000 UTC m=+704.384508938" watchObservedRunningTime="2025-12-17 08:32:48.839676194 +0000 UTC m=+704.384746146" Dec 17 08:32:51 crc kubenswrapper[4966]: I1217 08:32:51.811684 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc" event={"ID":"53e82fa5-de70-4ee3-85f2-727d9ae3c98a","Type":"ContainerStarted","Data":"8a3c2d24544861726a80b537377970e2b1efd08cdebd6a79978233b55b5d1832"} Dec 17 08:32:51 crc kubenswrapper[4966]: I1217 08:32:51.837077 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-92lfc" podStartSLOduration=1.817612137 podStartE2EDuration="8.83705284s" podCreationTimestamp="2025-12-17 08:32:43 +0000 UTC" firstStartedPulling="2025-12-17 08:32:44.355518402 +0000 UTC m=+699.900588344" lastFinishedPulling="2025-12-17 08:32:51.374959105 +0000 UTC m=+706.920029047" observedRunningTime="2025-12-17 08:32:51.832228845 +0000 UTC m=+707.377298817" watchObservedRunningTime="2025-12-17 08:32:51.83705284 +0000 UTC m=+707.382122792" Dec 17 08:32:54 crc kubenswrapper[4966]: I1217 08:32:54.352716 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-j2f2z" Dec 17 08:32:54 crc kubenswrapper[4966]: I1217 08:32:54.415633 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:54 crc kubenswrapper[4966]: I1217 08:32:54.415755 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:54 crc kubenswrapper[4966]: I1217 08:32:54.420791 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:54 crc kubenswrapper[4966]: I1217 08:32:54.837715 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-86cc5994cc-45pt9" Dec 17 08:32:54 crc kubenswrapper[4966]: I1217 08:32:54.891789 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-b998v"] Dec 17 08:33:04 crc kubenswrapper[4966]: I1217 08:33:04.607031 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-f8fb84555-2wr4k" Dec 17 08:33:18 crc kubenswrapper[4966]: I1217 08:33:18.946830 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg"] Dec 17 08:33:18 crc kubenswrapper[4966]: I1217 08:33:18.948448 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:18 crc kubenswrapper[4966]: I1217 08:33:18.952307 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 17 08:33:18 crc kubenswrapper[4966]: I1217 08:33:18.955780 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg"] Dec 17 08:33:18 crc kubenswrapper[4966]: I1217 08:33:18.993501 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:18 crc kubenswrapper[4966]: I1217 08:33:18.993727 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jbbv\" (UniqueName: \"kubernetes.io/projected/e5b6baa7-494c-418c-a4e8-23a7adbb8048-kube-api-access-2jbbv\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:18 crc kubenswrapper[4966]: I1217 08:33:18.993789 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:19 crc kubenswrapper[4966]: I1217 08:33:19.094779 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:19 crc kubenswrapper[4966]: I1217 08:33:19.094848 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:19 crc kubenswrapper[4966]: I1217 08:33:19.094948 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jbbv\" (UniqueName: \"kubernetes.io/projected/e5b6baa7-494c-418c-a4e8-23a7adbb8048-kube-api-access-2jbbv\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:19 crc kubenswrapper[4966]: I1217 08:33:19.095414 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:19 crc kubenswrapper[4966]: I1217 08:33:19.095653 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:19 crc kubenswrapper[4966]: I1217 08:33:19.112911 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jbbv\" (UniqueName: \"kubernetes.io/projected/e5b6baa7-494c-418c-a4e8-23a7adbb8048-kube-api-access-2jbbv\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:19 crc kubenswrapper[4966]: I1217 08:33:19.265222 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:19 crc kubenswrapper[4966]: I1217 08:33:19.746135 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg"] Dec 17 08:33:19 crc kubenswrapper[4966]: I1217 08:33:19.939236 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-b998v" podUID="992684e8-e50b-44e7-b30c-7169f4877695" containerName="console" containerID="cri-o://1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232" gracePeriod=15 Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.015628 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" event={"ID":"e5b6baa7-494c-418c-a4e8-23a7adbb8048","Type":"ContainerStarted","Data":"856d77da2c53aae0bafbccc784c6d192fd314852ffae7e73f574c5d1851ca65f"} Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.015677 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" event={"ID":"e5b6baa7-494c-418c-a4e8-23a7adbb8048","Type":"ContainerStarted","Data":"0f0d9a5b2535f6b5a8e3a237f0030fce5d5c5491a9d7946fe9652e89b79c4d8f"} Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.427192 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-b998v_992684e8-e50b-44e7-b30c-7169f4877695/console/0.log" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.427251 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.522197 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-serving-cert\") pod \"992684e8-e50b-44e7-b30c-7169f4877695\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.522279 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-service-ca\") pod \"992684e8-e50b-44e7-b30c-7169f4877695\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.522311 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-console-config\") pod \"992684e8-e50b-44e7-b30c-7169f4877695\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.522334 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcqx5\" (UniqueName: \"kubernetes.io/projected/992684e8-e50b-44e7-b30c-7169f4877695-kube-api-access-gcqx5\") pod \"992684e8-e50b-44e7-b30c-7169f4877695\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.522356 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-oauth-serving-cert\") pod \"992684e8-e50b-44e7-b30c-7169f4877695\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.522382 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-oauth-config\") pod \"992684e8-e50b-44e7-b30c-7169f4877695\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.522437 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-trusted-ca-bundle\") pod \"992684e8-e50b-44e7-b30c-7169f4877695\" (UID: \"992684e8-e50b-44e7-b30c-7169f4877695\") " Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.523792 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-console-config" (OuterVolumeSpecName: "console-config") pod "992684e8-e50b-44e7-b30c-7169f4877695" (UID: "992684e8-e50b-44e7-b30c-7169f4877695"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.524080 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "992684e8-e50b-44e7-b30c-7169f4877695" (UID: "992684e8-e50b-44e7-b30c-7169f4877695"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.524600 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-service-ca" (OuterVolumeSpecName: "service-ca") pod "992684e8-e50b-44e7-b30c-7169f4877695" (UID: "992684e8-e50b-44e7-b30c-7169f4877695"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.525103 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "992684e8-e50b-44e7-b30c-7169f4877695" (UID: "992684e8-e50b-44e7-b30c-7169f4877695"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.532124 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "992684e8-e50b-44e7-b30c-7169f4877695" (UID: "992684e8-e50b-44e7-b30c-7169f4877695"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.533300 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/992684e8-e50b-44e7-b30c-7169f4877695-kube-api-access-gcqx5" (OuterVolumeSpecName: "kube-api-access-gcqx5") pod "992684e8-e50b-44e7-b30c-7169f4877695" (UID: "992684e8-e50b-44e7-b30c-7169f4877695"). InnerVolumeSpecName "kube-api-access-gcqx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.548399 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "992684e8-e50b-44e7-b30c-7169f4877695" (UID: "992684e8-e50b-44e7-b30c-7169f4877695"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.623263 4966 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.623450 4966 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.623554 4966 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-service-ca\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.623611 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcqx5\" (UniqueName: \"kubernetes.io/projected/992684e8-e50b-44e7-b30c-7169f4877695-kube-api-access-gcqx5\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.623662 4966 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-console-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.623712 4966 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/992684e8-e50b-44e7-b30c-7169f4877695-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:20 crc kubenswrapper[4966]: I1217 08:33:20.623761 4966 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/992684e8-e50b-44e7-b30c-7169f4877695-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.023217 4966 generic.go:334] "Generic (PLEG): container finished" podID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerID="856d77da2c53aae0bafbccc784c6d192fd314852ffae7e73f574c5d1851ca65f" exitCode=0 Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.023283 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" event={"ID":"e5b6baa7-494c-418c-a4e8-23a7adbb8048","Type":"ContainerDied","Data":"856d77da2c53aae0bafbccc784c6d192fd314852ffae7e73f574c5d1851ca65f"} Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.026321 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-b998v_992684e8-e50b-44e7-b30c-7169f4877695/console/0.log" Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.026381 4966 generic.go:334] "Generic (PLEG): container finished" podID="992684e8-e50b-44e7-b30c-7169f4877695" containerID="1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232" exitCode=2 Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.026419 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-b998v" event={"ID":"992684e8-e50b-44e7-b30c-7169f4877695","Type":"ContainerDied","Data":"1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232"} Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.026454 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-b998v" event={"ID":"992684e8-e50b-44e7-b30c-7169f4877695","Type":"ContainerDied","Data":"4cfcdc71daf95981cede6a1817eda7862ce53866d79817de2661fd4e9c998bca"} Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.026481 4966 scope.go:117] "RemoveContainer" containerID="1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232" Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.026739 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-b998v" Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.056659 4966 scope.go:117] "RemoveContainer" containerID="1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232" Dec 17 08:33:21 crc kubenswrapper[4966]: E1217 08:33:21.057289 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232\": container with ID starting with 1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232 not found: ID does not exist" containerID="1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232" Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.057316 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232"} err="failed to get container status \"1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232\": rpc error: code = NotFound desc = could not find container \"1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232\": container with ID starting with 1b390a0ce11294f1177eb0916d93f4519e0eac73a2ea21b1ccb9ccf614a03232 not found: ID does not exist" Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.076026 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-b998v"] Dec 17 08:33:21 crc kubenswrapper[4966]: I1217 08:33:21.079808 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-b998v"] Dec 17 08:33:22 crc kubenswrapper[4966]: I1217 08:33:22.836856 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="992684e8-e50b-44e7-b30c-7169f4877695" path="/var/lib/kubelet/pods/992684e8-e50b-44e7-b30c-7169f4877695/volumes" Dec 17 08:33:23 crc kubenswrapper[4966]: I1217 08:33:23.051203 4966 generic.go:334] "Generic (PLEG): container finished" podID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerID="abca18c330b8cd4388f8b473abdbde152baca6d24e7a7842a5d449dd8e7bd0b8" exitCode=0 Dec 17 08:33:23 crc kubenswrapper[4966]: I1217 08:33:23.051275 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" event={"ID":"e5b6baa7-494c-418c-a4e8-23a7adbb8048","Type":"ContainerDied","Data":"abca18c330b8cd4388f8b473abdbde152baca6d24e7a7842a5d449dd8e7bd0b8"} Dec 17 08:33:24 crc kubenswrapper[4966]: I1217 08:33:24.058804 4966 generic.go:334] "Generic (PLEG): container finished" podID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerID="42b108adc9bffeee78b4606a3e6fbed5e2e3bb70a2d67c0ff7dffed6360c8d1b" exitCode=0 Dec 17 08:33:24 crc kubenswrapper[4966]: I1217 08:33:24.058993 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" event={"ID":"e5b6baa7-494c-418c-a4e8-23a7adbb8048","Type":"ContainerDied","Data":"42b108adc9bffeee78b4606a3e6fbed5e2e3bb70a2d67c0ff7dffed6360c8d1b"} Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.320043 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.489610 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jbbv\" (UniqueName: \"kubernetes.io/projected/e5b6baa7-494c-418c-a4e8-23a7adbb8048-kube-api-access-2jbbv\") pod \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.489747 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-util\") pod \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.489810 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-bundle\") pod \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\" (UID: \"e5b6baa7-494c-418c-a4e8-23a7adbb8048\") " Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.490915 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-bundle" (OuterVolumeSpecName: "bundle") pod "e5b6baa7-494c-418c-a4e8-23a7adbb8048" (UID: "e5b6baa7-494c-418c-a4e8-23a7adbb8048"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.491343 4966 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.499056 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5b6baa7-494c-418c-a4e8-23a7adbb8048-kube-api-access-2jbbv" (OuterVolumeSpecName: "kube-api-access-2jbbv") pod "e5b6baa7-494c-418c-a4e8-23a7adbb8048" (UID: "e5b6baa7-494c-418c-a4e8-23a7adbb8048"). InnerVolumeSpecName "kube-api-access-2jbbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.501382 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-util" (OuterVolumeSpecName: "util") pod "e5b6baa7-494c-418c-a4e8-23a7adbb8048" (UID: "e5b6baa7-494c-418c-a4e8-23a7adbb8048"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.592834 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jbbv\" (UniqueName: \"kubernetes.io/projected/e5b6baa7-494c-418c-a4e8-23a7adbb8048-kube-api-access-2jbbv\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:25 crc kubenswrapper[4966]: I1217 08:33:25.592901 4966 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e5b6baa7-494c-418c-a4e8-23a7adbb8048-util\") on node \"crc\" DevicePath \"\"" Dec 17 08:33:26 crc kubenswrapper[4966]: I1217 08:33:26.072683 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" event={"ID":"e5b6baa7-494c-418c-a4e8-23a7adbb8048","Type":"ContainerDied","Data":"0f0d9a5b2535f6b5a8e3a237f0030fce5d5c5491a9d7946fe9652e89b79c4d8f"} Dec 17 08:33:26 crc kubenswrapper[4966]: I1217 08:33:26.072758 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f0d9a5b2535f6b5a8e3a237f0030fce5d5c5491a9d7946fe9652e89b79c4d8f" Dec 17 08:33:26 crc kubenswrapper[4966]: I1217 08:33:26.072801 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4m8qwg" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.391774 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869"] Dec 17 08:33:34 crc kubenswrapper[4966]: E1217 08:33:34.392477 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerName="util" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.392490 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerName="util" Dec 17 08:33:34 crc kubenswrapper[4966]: E1217 08:33:34.392514 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992684e8-e50b-44e7-b30c-7169f4877695" containerName="console" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.392522 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="992684e8-e50b-44e7-b30c-7169f4877695" containerName="console" Dec 17 08:33:34 crc kubenswrapper[4966]: E1217 08:33:34.392534 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerName="pull" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.392541 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerName="pull" Dec 17 08:33:34 crc kubenswrapper[4966]: E1217 08:33:34.392550 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerName="extract" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.392556 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerName="extract" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.392669 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="992684e8-e50b-44e7-b30c-7169f4877695" containerName="console" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.392685 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5b6baa7-494c-418c-a4e8-23a7adbb8048" containerName="extract" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.393128 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.402408 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.402980 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.403185 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.403308 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-2g7vx" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.403408 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.471653 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869"] Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.509249 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42df3cf4-b96d-4674-be0b-8386748de493-webhook-cert\") pod \"metallb-operator-controller-manager-6bddf8b9f4-2h869\" (UID: \"42df3cf4-b96d-4674-be0b-8386748de493\") " pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.509319 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42df3cf4-b96d-4674-be0b-8386748de493-apiservice-cert\") pod \"metallb-operator-controller-manager-6bddf8b9f4-2h869\" (UID: \"42df3cf4-b96d-4674-be0b-8386748de493\") " pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.509349 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6tnm\" (UniqueName: \"kubernetes.io/projected/42df3cf4-b96d-4674-be0b-8386748de493-kube-api-access-q6tnm\") pod \"metallb-operator-controller-manager-6bddf8b9f4-2h869\" (UID: \"42df3cf4-b96d-4674-be0b-8386748de493\") " pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.610479 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42df3cf4-b96d-4674-be0b-8386748de493-webhook-cert\") pod \"metallb-operator-controller-manager-6bddf8b9f4-2h869\" (UID: \"42df3cf4-b96d-4674-be0b-8386748de493\") " pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.610760 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42df3cf4-b96d-4674-be0b-8386748de493-apiservice-cert\") pod \"metallb-operator-controller-manager-6bddf8b9f4-2h869\" (UID: \"42df3cf4-b96d-4674-be0b-8386748de493\") " pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.610852 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6tnm\" (UniqueName: \"kubernetes.io/projected/42df3cf4-b96d-4674-be0b-8386748de493-kube-api-access-q6tnm\") pod \"metallb-operator-controller-manager-6bddf8b9f4-2h869\" (UID: \"42df3cf4-b96d-4674-be0b-8386748de493\") " pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.619743 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/42df3cf4-b96d-4674-be0b-8386748de493-apiservice-cert\") pod \"metallb-operator-controller-manager-6bddf8b9f4-2h869\" (UID: \"42df3cf4-b96d-4674-be0b-8386748de493\") " pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.619786 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/42df3cf4-b96d-4674-be0b-8386748de493-webhook-cert\") pod \"metallb-operator-controller-manager-6bddf8b9f4-2h869\" (UID: \"42df3cf4-b96d-4674-be0b-8386748de493\") " pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.638757 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6tnm\" (UniqueName: \"kubernetes.io/projected/42df3cf4-b96d-4674-be0b-8386748de493-kube-api-access-q6tnm\") pod \"metallb-operator-controller-manager-6bddf8b9f4-2h869\" (UID: \"42df3cf4-b96d-4674-be0b-8386748de493\") " pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.742409 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.910702 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk"] Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.911963 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.929812 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.930065 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 17 08:33:34 crc kubenswrapper[4966]: I1217 08:33:34.930237 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-s9kpv" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.015053 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7588e41-5206-4295-9260-991ef73c9551-apiservice-cert\") pod \"metallb-operator-webhook-server-fff4478f-4xjqk\" (UID: \"a7588e41-5206-4295-9260-991ef73c9551\") " pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.015092 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4fz8\" (UniqueName: \"kubernetes.io/projected/a7588e41-5206-4295-9260-991ef73c9551-kube-api-access-z4fz8\") pod \"metallb-operator-webhook-server-fff4478f-4xjqk\" (UID: \"a7588e41-5206-4295-9260-991ef73c9551\") " pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.015143 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7588e41-5206-4295-9260-991ef73c9551-webhook-cert\") pod \"metallb-operator-webhook-server-fff4478f-4xjqk\" (UID: \"a7588e41-5206-4295-9260-991ef73c9551\") " pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.027259 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk"] Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.116321 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7588e41-5206-4295-9260-991ef73c9551-webhook-cert\") pod \"metallb-operator-webhook-server-fff4478f-4xjqk\" (UID: \"a7588e41-5206-4295-9260-991ef73c9551\") " pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.116438 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7588e41-5206-4295-9260-991ef73c9551-apiservice-cert\") pod \"metallb-operator-webhook-server-fff4478f-4xjqk\" (UID: \"a7588e41-5206-4295-9260-991ef73c9551\") " pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.116466 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4fz8\" (UniqueName: \"kubernetes.io/projected/a7588e41-5206-4295-9260-991ef73c9551-kube-api-access-z4fz8\") pod \"metallb-operator-webhook-server-fff4478f-4xjqk\" (UID: \"a7588e41-5206-4295-9260-991ef73c9551\") " pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.121249 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7588e41-5206-4295-9260-991ef73c9551-apiservice-cert\") pod \"metallb-operator-webhook-server-fff4478f-4xjqk\" (UID: \"a7588e41-5206-4295-9260-991ef73c9551\") " pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.128949 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7588e41-5206-4295-9260-991ef73c9551-webhook-cert\") pod \"metallb-operator-webhook-server-fff4478f-4xjqk\" (UID: \"a7588e41-5206-4295-9260-991ef73c9551\") " pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.143227 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4fz8\" (UniqueName: \"kubernetes.io/projected/a7588e41-5206-4295-9260-991ef73c9551-kube-api-access-z4fz8\") pod \"metallb-operator-webhook-server-fff4478f-4xjqk\" (UID: \"a7588e41-5206-4295-9260-991ef73c9551\") " pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.172956 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869"] Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.286245 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:35 crc kubenswrapper[4966]: I1217 08:33:35.666529 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk"] Dec 17 08:33:36 crc kubenswrapper[4966]: I1217 08:33:36.164855 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" event={"ID":"a7588e41-5206-4295-9260-991ef73c9551","Type":"ContainerStarted","Data":"b5afac8f83fc766c81db41e6ca5f155bfc7881cae352255fccdd09f901cec370"} Dec 17 08:33:36 crc kubenswrapper[4966]: I1217 08:33:36.166593 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" event={"ID":"42df3cf4-b96d-4674-be0b-8386748de493","Type":"ContainerStarted","Data":"0a49de8bf15de24837a8147d902f6c4a880550ec282980db682663488eb85229"} Dec 17 08:33:43 crc kubenswrapper[4966]: I1217 08:33:43.215155 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" event={"ID":"42df3cf4-b96d-4674-be0b-8386748de493","Type":"ContainerStarted","Data":"cbe045dd153bb8c35e571d9e31eb8bbc4f0e46af6143ea102dba9e31be2cfd29"} Dec 17 08:33:43 crc kubenswrapper[4966]: I1217 08:33:43.215642 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:33:43 crc kubenswrapper[4966]: I1217 08:33:43.217526 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" event={"ID":"a7588e41-5206-4295-9260-991ef73c9551","Type":"ContainerStarted","Data":"e0db9e0b3bba060e1026e4c194819daf18d7cb2e7b18b1573ca3804084860b97"} Dec 17 08:33:43 crc kubenswrapper[4966]: I1217 08:33:43.217649 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:43 crc kubenswrapper[4966]: I1217 08:33:43.243486 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" podStartSLOduration=1.9088863109999998 podStartE2EDuration="9.243463279s" podCreationTimestamp="2025-12-17 08:33:34 +0000 UTC" firstStartedPulling="2025-12-17 08:33:35.192945151 +0000 UTC m=+750.738015093" lastFinishedPulling="2025-12-17 08:33:42.527522129 +0000 UTC m=+758.072592061" observedRunningTime="2025-12-17 08:33:43.235342643 +0000 UTC m=+758.780412585" watchObservedRunningTime="2025-12-17 08:33:43.243463279 +0000 UTC m=+758.788533241" Dec 17 08:33:43 crc kubenswrapper[4966]: I1217 08:33:43.265181 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" podStartSLOduration=2.367499259 podStartE2EDuration="9.265161127s" podCreationTimestamp="2025-12-17 08:33:34 +0000 UTC" firstStartedPulling="2025-12-17 08:33:35.672016211 +0000 UTC m=+751.217086153" lastFinishedPulling="2025-12-17 08:33:42.569678079 +0000 UTC m=+758.114748021" observedRunningTime="2025-12-17 08:33:43.261294789 +0000 UTC m=+758.806364731" watchObservedRunningTime="2025-12-17 08:33:43.265161127 +0000 UTC m=+758.810231069" Dec 17 08:33:55 crc kubenswrapper[4966]: I1217 08:33:55.295423 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-fff4478f-4xjqk" Dec 17 08:33:57 crc kubenswrapper[4966]: I1217 08:33:57.345996 4966 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 17 08:34:14 crc kubenswrapper[4966]: I1217 08:34:14.745435 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6bddf8b9f4-2h869" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.427121 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-gvbzk"] Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.429481 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.432693 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-x7qmw" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.433262 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.433677 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.442412 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m"] Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.443567 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.448116 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.449057 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m"] Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.513811 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-frr-sockets\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.513907 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-reloader\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.513979 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8vf2\" (UniqueName: \"kubernetes.io/projected/fcede706-7499-4c33-b80e-e1a13b7474dd-kube-api-access-z8vf2\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.514013 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcede706-7499-4c33-b80e-e1a13b7474dd-metrics-certs\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.514044 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-metrics\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.515823 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6vrz\" (UniqueName: \"kubernetes.io/projected/9f0de3ea-2dfd-411b-a2a6-419398ae4739-kube-api-access-j6vrz\") pod \"frr-k8s-webhook-server-7784b6fcf-h9g4m\" (UID: \"9f0de3ea-2dfd-411b-a2a6-419398ae4739\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.515889 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f0de3ea-2dfd-411b-a2a6-419398ae4739-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-h9g4m\" (UID: \"9f0de3ea-2dfd-411b-a2a6-419398ae4739\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.515964 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-frr-conf\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.516019 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fcede706-7499-4c33-b80e-e1a13b7474dd-frr-startup\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.554370 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-pllwn"] Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.555506 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-8mc27"] Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.556488 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.557433 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.562383 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.562459 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-27bpq" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.562585 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.562780 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.567459 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.569838 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-pllwn"] Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.617344 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b77af9c-5db2-4792-8de6-f882bc6656d4-metrics-certs\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.617552 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8vf2\" (UniqueName: \"kubernetes.io/projected/fcede706-7499-4c33-b80e-e1a13b7474dd-kube-api-access-z8vf2\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.617650 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-metrics-certs\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.617751 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcede706-7499-4c33-b80e-e1a13b7474dd-metrics-certs\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.617834 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/761f12b2-ba49-4859-9d5e-7903923c4ac9-metallb-excludel2\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.617924 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618002 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-metrics\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618072 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6vrz\" (UniqueName: \"kubernetes.io/projected/9f0de3ea-2dfd-411b-a2a6-419398ae4739-kube-api-access-j6vrz\") pod \"frr-k8s-webhook-server-7784b6fcf-h9g4m\" (UID: \"9f0de3ea-2dfd-411b-a2a6-419398ae4739\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618139 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ktbx\" (UniqueName: \"kubernetes.io/projected/761f12b2-ba49-4859-9d5e-7903923c4ac9-kube-api-access-4ktbx\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618233 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f0de3ea-2dfd-411b-a2a6-419398ae4739-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-h9g4m\" (UID: \"9f0de3ea-2dfd-411b-a2a6-419398ae4739\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618329 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w9gm\" (UniqueName: \"kubernetes.io/projected/3b77af9c-5db2-4792-8de6-f882bc6656d4-kube-api-access-4w9gm\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618414 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b77af9c-5db2-4792-8de6-f882bc6656d4-cert\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618489 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-frr-conf\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618575 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fcede706-7499-4c33-b80e-e1a13b7474dd-frr-startup\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618643 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-frr-sockets\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.618722 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-reloader\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.619149 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-reloader\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: E1217 08:34:15.619513 4966 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 17 08:34:15 crc kubenswrapper[4966]: E1217 08:34:15.619618 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fcede706-7499-4c33-b80e-e1a13b7474dd-metrics-certs podName:fcede706-7499-4c33-b80e-e1a13b7474dd nodeName:}" failed. No retries permitted until 2025-12-17 08:34:16.119603737 +0000 UTC m=+791.664673679 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fcede706-7499-4c33-b80e-e1a13b7474dd-metrics-certs") pod "frr-k8s-gvbzk" (UID: "fcede706-7499-4c33-b80e-e1a13b7474dd") : secret "frr-k8s-certs-secret" not found Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.620014 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-metrics\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: E1217 08:34:15.620266 4966 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 17 08:34:15 crc kubenswrapper[4966]: E1217 08:34:15.620370 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f0de3ea-2dfd-411b-a2a6-419398ae4739-cert podName:9f0de3ea-2dfd-411b-a2a6-419398ae4739 nodeName:}" failed. No retries permitted until 2025-12-17 08:34:16.120360978 +0000 UTC m=+791.665430920 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9f0de3ea-2dfd-411b-a2a6-419398ae4739-cert") pod "frr-k8s-webhook-server-7784b6fcf-h9g4m" (UID: "9f0de3ea-2dfd-411b-a2a6-419398ae4739") : secret "frr-k8s-webhook-server-cert" not found Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.620628 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-frr-conf\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.621408 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fcede706-7499-4c33-b80e-e1a13b7474dd-frr-startup\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.621676 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fcede706-7499-4c33-b80e-e1a13b7474dd-frr-sockets\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.643718 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8vf2\" (UniqueName: \"kubernetes.io/projected/fcede706-7499-4c33-b80e-e1a13b7474dd-kube-api-access-z8vf2\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.660981 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6vrz\" (UniqueName: \"kubernetes.io/projected/9f0de3ea-2dfd-411b-a2a6-419398ae4739-kube-api-access-j6vrz\") pod \"frr-k8s-webhook-server-7784b6fcf-h9g4m\" (UID: \"9f0de3ea-2dfd-411b-a2a6-419398ae4739\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.720465 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b77af9c-5db2-4792-8de6-f882bc6656d4-metrics-certs\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:15 crc kubenswrapper[4966]: E1217 08:34:15.720757 4966 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 17 08:34:15 crc kubenswrapper[4966]: E1217 08:34:15.720851 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3b77af9c-5db2-4792-8de6-f882bc6656d4-metrics-certs podName:3b77af9c-5db2-4792-8de6-f882bc6656d4 nodeName:}" failed. No retries permitted until 2025-12-17 08:34:16.22082821 +0000 UTC m=+791.765898152 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3b77af9c-5db2-4792-8de6-f882bc6656d4-metrics-certs") pod "controller-5bddd4b946-pllwn" (UID: "3b77af9c-5db2-4792-8de6-f882bc6656d4") : secret "controller-certs-secret" not found Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.720981 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-metrics-certs\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.721103 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/761f12b2-ba49-4859-9d5e-7903923c4ac9-metallb-excludel2\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.721202 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: E1217 08:34:15.721313 4966 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 17 08:34:15 crc kubenswrapper[4966]: E1217 08:34:15.721359 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist podName:761f12b2-ba49-4859-9d5e-7903923c4ac9 nodeName:}" failed. No retries permitted until 2025-12-17 08:34:16.221345644 +0000 UTC m=+791.766415656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist") pod "speaker-8mc27" (UID: "761f12b2-ba49-4859-9d5e-7903923c4ac9") : secret "metallb-memberlist" not found Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.721323 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ktbx\" (UniqueName: \"kubernetes.io/projected/761f12b2-ba49-4859-9d5e-7903923c4ac9-kube-api-access-4ktbx\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.721581 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w9gm\" (UniqueName: \"kubernetes.io/projected/3b77af9c-5db2-4792-8de6-f882bc6656d4-kube-api-access-4w9gm\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.721688 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b77af9c-5db2-4792-8de6-f882bc6656d4-cert\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.721825 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/761f12b2-ba49-4859-9d5e-7903923c4ac9-metallb-excludel2\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.724271 4966 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.725389 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-metrics-certs\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.737322 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b77af9c-5db2-4792-8de6-f882bc6656d4-cert\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.742770 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ktbx\" (UniqueName: \"kubernetes.io/projected/761f12b2-ba49-4859-9d5e-7903923c4ac9-kube-api-access-4ktbx\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:15 crc kubenswrapper[4966]: I1217 08:34:15.748532 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w9gm\" (UniqueName: \"kubernetes.io/projected/3b77af9c-5db2-4792-8de6-f882bc6656d4-kube-api-access-4w9gm\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.126800 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcede706-7499-4c33-b80e-e1a13b7474dd-metrics-certs\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.127192 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f0de3ea-2dfd-411b-a2a6-419398ae4739-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-h9g4m\" (UID: \"9f0de3ea-2dfd-411b-a2a6-419398ae4739\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.130368 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f0de3ea-2dfd-411b-a2a6-419398ae4739-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-h9g4m\" (UID: \"9f0de3ea-2dfd-411b-a2a6-419398ae4739\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.130475 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fcede706-7499-4c33-b80e-e1a13b7474dd-metrics-certs\") pod \"frr-k8s-gvbzk\" (UID: \"fcede706-7499-4c33-b80e-e1a13b7474dd\") " pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.229014 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b77af9c-5db2-4792-8de6-f882bc6656d4-metrics-certs\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.229279 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:16 crc kubenswrapper[4966]: E1217 08:34:16.229403 4966 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 17 08:34:16 crc kubenswrapper[4966]: E1217 08:34:16.229481 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist podName:761f12b2-ba49-4859-9d5e-7903923c4ac9 nodeName:}" failed. No retries permitted until 2025-12-17 08:34:17.229465208 +0000 UTC m=+792.774535150 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist") pod "speaker-8mc27" (UID: "761f12b2-ba49-4859-9d5e-7903923c4ac9") : secret "metallb-memberlist" not found Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.233546 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b77af9c-5db2-4792-8de6-f882bc6656d4-metrics-certs\") pod \"controller-5bddd4b946-pllwn\" (UID: \"3b77af9c-5db2-4792-8de6-f882bc6656d4\") " pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.356909 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.366736 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.492995 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.759218 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m"] Dec 17 08:34:16 crc kubenswrapper[4966]: W1217 08:34:16.764566 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f0de3ea_2dfd_411b_a2a6_419398ae4739.slice/crio-f40f68e1595b238bbc5c7eb29d4925005cc29c9d2c8450fde79f988a9628b58c WatchSource:0}: Error finding container f40f68e1595b238bbc5c7eb29d4925005cc29c9d2c8450fde79f988a9628b58c: Status 404 returned error can't find the container with id f40f68e1595b238bbc5c7eb29d4925005cc29c9d2c8450fde79f988a9628b58c Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.807347 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.807393 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:34:16 crc kubenswrapper[4966]: I1217 08:34:16.890407 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-pllwn"] Dec 17 08:34:17 crc kubenswrapper[4966]: I1217 08:34:17.245942 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:17 crc kubenswrapper[4966]: E1217 08:34:17.246218 4966 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 17 08:34:17 crc kubenswrapper[4966]: E1217 08:34:17.246290 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist podName:761f12b2-ba49-4859-9d5e-7903923c4ac9 nodeName:}" failed. No retries permitted until 2025-12-17 08:34:19.24627391 +0000 UTC m=+794.791343852 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist") pod "speaker-8mc27" (UID: "761f12b2-ba49-4859-9d5e-7903923c4ac9") : secret "metallb-memberlist" not found Dec 17 08:34:17 crc kubenswrapper[4966]: I1217 08:34:17.446481 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-pllwn" event={"ID":"3b77af9c-5db2-4792-8de6-f882bc6656d4","Type":"ContainerStarted","Data":"1528b2d46249a97e746768a662b0fc2075474410ecf67ce01c02ba9b5975f4c2"} Dec 17 08:34:17 crc kubenswrapper[4966]: I1217 08:34:17.446532 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-pllwn" event={"ID":"3b77af9c-5db2-4792-8de6-f882bc6656d4","Type":"ContainerStarted","Data":"9f6d016c29e91dbab5d960a4f24d5199c13df7ad9ef17f6cf40c537fe117ab26"} Dec 17 08:34:17 crc kubenswrapper[4966]: I1217 08:34:17.446547 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-pllwn" event={"ID":"3b77af9c-5db2-4792-8de6-f882bc6656d4","Type":"ContainerStarted","Data":"b09160df6cf73ef94fb8e25c6998447aeeade8f91e91d1a1ec306f3df3064c02"} Dec 17 08:34:17 crc kubenswrapper[4966]: I1217 08:34:17.446604 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:17 crc kubenswrapper[4966]: I1217 08:34:17.448323 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" event={"ID":"9f0de3ea-2dfd-411b-a2a6-419398ae4739","Type":"ContainerStarted","Data":"f40f68e1595b238bbc5c7eb29d4925005cc29c9d2c8450fde79f988a9628b58c"} Dec 17 08:34:17 crc kubenswrapper[4966]: I1217 08:34:17.449366 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerStarted","Data":"354c19b51e68f9621a1ddd2feb73769fbee51208fe81d1678d11fbb2311a97a7"} Dec 17 08:34:19 crc kubenswrapper[4966]: I1217 08:34:19.278337 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:19 crc kubenswrapper[4966]: I1217 08:34:19.290412 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/761f12b2-ba49-4859-9d5e-7903923c4ac9-memberlist\") pod \"speaker-8mc27\" (UID: \"761f12b2-ba49-4859-9d5e-7903923c4ac9\") " pod="metallb-system/speaker-8mc27" Dec 17 08:34:19 crc kubenswrapper[4966]: I1217 08:34:19.484098 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8mc27" Dec 17 08:34:19 crc kubenswrapper[4966]: W1217 08:34:19.619096 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod761f12b2_ba49_4859_9d5e_7903923c4ac9.slice/crio-e664478b116449630b1a38006f632c37c0e2ce510de632c1098e8e2ea0987440 WatchSource:0}: Error finding container e664478b116449630b1a38006f632c37c0e2ce510de632c1098e8e2ea0987440: Status 404 returned error can't find the container with id e664478b116449630b1a38006f632c37c0e2ce510de632c1098e8e2ea0987440 Dec 17 08:34:20 crc kubenswrapper[4966]: I1217 08:34:20.483228 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8mc27" event={"ID":"761f12b2-ba49-4859-9d5e-7903923c4ac9","Type":"ContainerStarted","Data":"26b2acbeaaecc27030f9619665b6ab3d6512fe207259cc53c76f10a4c260a2b1"} Dec 17 08:34:20 crc kubenswrapper[4966]: I1217 08:34:20.483593 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8mc27" event={"ID":"761f12b2-ba49-4859-9d5e-7903923c4ac9","Type":"ContainerStarted","Data":"c1f810df7871787c54f9ad691d89da530b310772a8c0772f4ea3c42575f1f79b"} Dec 17 08:34:20 crc kubenswrapper[4966]: I1217 08:34:20.483611 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8mc27" event={"ID":"761f12b2-ba49-4859-9d5e-7903923c4ac9","Type":"ContainerStarted","Data":"e664478b116449630b1a38006f632c37c0e2ce510de632c1098e8e2ea0987440"} Dec 17 08:34:20 crc kubenswrapper[4966]: I1217 08:34:20.483819 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-8mc27" Dec 17 08:34:20 crc kubenswrapper[4966]: I1217 08:34:20.502762 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-pllwn" podStartSLOduration=5.5027438140000005 podStartE2EDuration="5.502743814s" podCreationTimestamp="2025-12-17 08:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:34:17.470373554 +0000 UTC m=+793.015443496" watchObservedRunningTime="2025-12-17 08:34:20.502743814 +0000 UTC m=+796.047813756" Dec 17 08:34:24 crc kubenswrapper[4966]: I1217 08:34:24.859668 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-8mc27" podStartSLOduration=9.858662795 podStartE2EDuration="9.858662795s" podCreationTimestamp="2025-12-17 08:34:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:34:20.50436863 +0000 UTC m=+796.049438572" watchObservedRunningTime="2025-12-17 08:34:24.858662795 +0000 UTC m=+800.403732737" Dec 17 08:34:26 crc kubenswrapper[4966]: I1217 08:34:26.523655 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" event={"ID":"9f0de3ea-2dfd-411b-a2a6-419398ae4739","Type":"ContainerStarted","Data":"af8316e3e948bfb94fcf21fd5b46232f0178c5f09c5f47827043095281e0d94e"} Dec 17 08:34:26 crc kubenswrapper[4966]: I1217 08:34:26.524174 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:26 crc kubenswrapper[4966]: I1217 08:34:26.524891 4966 generic.go:334] "Generic (PLEG): container finished" podID="fcede706-7499-4c33-b80e-e1a13b7474dd" containerID="6b480705fae1499be4a5ce3258767ae29d1ad55b21ac45b4ca8a35f6a96b989f" exitCode=0 Dec 17 08:34:26 crc kubenswrapper[4966]: I1217 08:34:26.524935 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerDied","Data":"6b480705fae1499be4a5ce3258767ae29d1ad55b21ac45b4ca8a35f6a96b989f"} Dec 17 08:34:26 crc kubenswrapper[4966]: I1217 08:34:26.544289 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" podStartSLOduration=2.6610714619999998 podStartE2EDuration="11.544267478s" podCreationTimestamp="2025-12-17 08:34:15 +0000 UTC" firstStartedPulling="2025-12-17 08:34:16.76753207 +0000 UTC m=+792.312602012" lastFinishedPulling="2025-12-17 08:34:25.650728066 +0000 UTC m=+801.195798028" observedRunningTime="2025-12-17 08:34:26.540294317 +0000 UTC m=+802.085364259" watchObservedRunningTime="2025-12-17 08:34:26.544267478 +0000 UTC m=+802.089337420" Dec 17 08:34:27 crc kubenswrapper[4966]: I1217 08:34:27.534543 4966 generic.go:334] "Generic (PLEG): container finished" podID="fcede706-7499-4c33-b80e-e1a13b7474dd" containerID="fff8f42f16cd2db1447520c65dc9817e2f501e4c5f9ebb0db21c4fe12d225ad3" exitCode=0 Dec 17 08:34:27 crc kubenswrapper[4966]: I1217 08:34:27.534647 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerDied","Data":"fff8f42f16cd2db1447520c65dc9817e2f501e4c5f9ebb0db21c4fe12d225ad3"} Dec 17 08:34:28 crc kubenswrapper[4966]: I1217 08:34:28.542946 4966 generic.go:334] "Generic (PLEG): container finished" podID="fcede706-7499-4c33-b80e-e1a13b7474dd" containerID="7a9dcabe558d1d8b7d164c41440cad2aee4e5c42e395c73178c8f5722aa4711e" exitCode=0 Dec 17 08:34:28 crc kubenswrapper[4966]: I1217 08:34:28.542994 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerDied","Data":"7a9dcabe558d1d8b7d164c41440cad2aee4e5c42e395c73178c8f5722aa4711e"} Dec 17 08:34:29 crc kubenswrapper[4966]: I1217 08:34:29.488398 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-8mc27" Dec 17 08:34:29 crc kubenswrapper[4966]: I1217 08:34:29.561468 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerStarted","Data":"a156d9bc34792b27e05c91e41ab1d87583056e3b434ce7c74bf7395cb5255f51"} Dec 17 08:34:29 crc kubenswrapper[4966]: I1217 08:34:29.561506 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerStarted","Data":"5ff7c04f9da129106ee64ce2e408c948faeb365ccda625a02c9a58f433af556d"} Dec 17 08:34:29 crc kubenswrapper[4966]: I1217 08:34:29.561516 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerStarted","Data":"74de93cf1d150d4639b38e005fddeeac76c68edbc77768e1532c53b1ebc9fb79"} Dec 17 08:34:29 crc kubenswrapper[4966]: I1217 08:34:29.561525 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerStarted","Data":"eb9848cf6c986a6dae35923a6423d35595cb0b2c0c46ee9c2a35789685a93383"} Dec 17 08:34:29 crc kubenswrapper[4966]: I1217 08:34:29.561533 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerStarted","Data":"21e17623a6012d45c385be7fbd74cc5d584600f9b117fed3d667990302a8df60"} Dec 17 08:34:30 crc kubenswrapper[4966]: I1217 08:34:30.570446 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gvbzk" event={"ID":"fcede706-7499-4c33-b80e-e1a13b7474dd","Type":"ContainerStarted","Data":"c9d833fb729601334126fcffc41ff93ec6f89aa90e0e3bee1db7c7095647ee0a"} Dec 17 08:34:30 crc kubenswrapper[4966]: I1217 08:34:30.570683 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:30 crc kubenswrapper[4966]: I1217 08:34:30.596561 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-gvbzk" podStartSLOduration=7.12987599 podStartE2EDuration="15.596540508s" podCreationTimestamp="2025-12-17 08:34:15 +0000 UTC" firstStartedPulling="2025-12-17 08:34:17.193104512 +0000 UTC m=+792.738174454" lastFinishedPulling="2025-12-17 08:34:25.65976899 +0000 UTC m=+801.204838972" observedRunningTime="2025-12-17 08:34:30.590634393 +0000 UTC m=+806.135704335" watchObservedRunningTime="2025-12-17 08:34:30.596540508 +0000 UTC m=+806.141610450" Dec 17 08:34:31 crc kubenswrapper[4966]: I1217 08:34:31.357925 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:31 crc kubenswrapper[4966]: I1217 08:34:31.395964 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.262658 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rzhjf"] Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.263425 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rzhjf" Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.272305 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-sf79f" Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.272493 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.272642 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.279435 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rzhjf"] Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.383227 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8rpv\" (UniqueName: \"kubernetes.io/projected/ac78d231-7df6-4e74-aad8-166ba49f6176-kube-api-access-t8rpv\") pod \"openstack-operator-index-rzhjf\" (UID: \"ac78d231-7df6-4e74-aad8-166ba49f6176\") " pod="openstack-operators/openstack-operator-index-rzhjf" Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.485021 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8rpv\" (UniqueName: \"kubernetes.io/projected/ac78d231-7df6-4e74-aad8-166ba49f6176-kube-api-access-t8rpv\") pod \"openstack-operator-index-rzhjf\" (UID: \"ac78d231-7df6-4e74-aad8-166ba49f6176\") " pod="openstack-operators/openstack-operator-index-rzhjf" Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.505297 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8rpv\" (UniqueName: \"kubernetes.io/projected/ac78d231-7df6-4e74-aad8-166ba49f6176-kube-api-access-t8rpv\") pod \"openstack-operator-index-rzhjf\" (UID: \"ac78d231-7df6-4e74-aad8-166ba49f6176\") " pod="openstack-operators/openstack-operator-index-rzhjf" Dec 17 08:34:32 crc kubenswrapper[4966]: I1217 08:34:32.581204 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rzhjf" Dec 17 08:34:33 crc kubenswrapper[4966]: I1217 08:34:33.060647 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rzhjf"] Dec 17 08:34:33 crc kubenswrapper[4966]: W1217 08:34:33.063831 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac78d231_7df6_4e74_aad8_166ba49f6176.slice/crio-1b0a3597564b7d5eeca8e9ca0dfc9b0a31d35a40a8efb6943529a4634d87540c WatchSource:0}: Error finding container 1b0a3597564b7d5eeca8e9ca0dfc9b0a31d35a40a8efb6943529a4634d87540c: Status 404 returned error can't find the container with id 1b0a3597564b7d5eeca8e9ca0dfc9b0a31d35a40a8efb6943529a4634d87540c Dec 17 08:34:33 crc kubenswrapper[4966]: I1217 08:34:33.591490 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rzhjf" event={"ID":"ac78d231-7df6-4e74-aad8-166ba49f6176","Type":"ContainerStarted","Data":"1b0a3597564b7d5eeca8e9ca0dfc9b0a31d35a40a8efb6943529a4634d87540c"} Dec 17 08:34:35 crc kubenswrapper[4966]: I1217 08:34:35.418916 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rzhjf"] Dec 17 08:34:36 crc kubenswrapper[4966]: I1217 08:34:36.028267 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5tcwp"] Dec 17 08:34:36 crc kubenswrapper[4966]: I1217 08:34:36.029440 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5tcwp" Dec 17 08:34:36 crc kubenswrapper[4966]: I1217 08:34:36.042330 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5tcwp"] Dec 17 08:34:36 crc kubenswrapper[4966]: I1217 08:34:36.134052 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvtk9\" (UniqueName: \"kubernetes.io/projected/872aa43a-f987-408d-a4c8-048df99f2244-kube-api-access-xvtk9\") pod \"openstack-operator-index-5tcwp\" (UID: \"872aa43a-f987-408d-a4c8-048df99f2244\") " pod="openstack-operators/openstack-operator-index-5tcwp" Dec 17 08:34:36 crc kubenswrapper[4966]: I1217 08:34:36.235234 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvtk9\" (UniqueName: \"kubernetes.io/projected/872aa43a-f987-408d-a4c8-048df99f2244-kube-api-access-xvtk9\") pod \"openstack-operator-index-5tcwp\" (UID: \"872aa43a-f987-408d-a4c8-048df99f2244\") " pod="openstack-operators/openstack-operator-index-5tcwp" Dec 17 08:34:36 crc kubenswrapper[4966]: I1217 08:34:36.257163 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvtk9\" (UniqueName: \"kubernetes.io/projected/872aa43a-f987-408d-a4c8-048df99f2244-kube-api-access-xvtk9\") pod \"openstack-operator-index-5tcwp\" (UID: \"872aa43a-f987-408d-a4c8-048df99f2244\") " pod="openstack-operators/openstack-operator-index-5tcwp" Dec 17 08:34:36 crc kubenswrapper[4966]: I1217 08:34:36.352976 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5tcwp" Dec 17 08:34:36 crc kubenswrapper[4966]: I1217 08:34:36.374639 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-h9g4m" Dec 17 08:34:36 crc kubenswrapper[4966]: I1217 08:34:36.500091 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-pllwn" Dec 17 08:34:37 crc kubenswrapper[4966]: I1217 08:34:37.906431 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5tcwp"] Dec 17 08:34:38 crc kubenswrapper[4966]: I1217 08:34:38.624746 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5tcwp" event={"ID":"872aa43a-f987-408d-a4c8-048df99f2244","Type":"ContainerStarted","Data":"a10bbe031d57954babcf68f7d3dd74fc16a4d059d959383555afc5fb53d4587a"} Dec 17 08:34:38 crc kubenswrapper[4966]: I1217 08:34:38.624800 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5tcwp" event={"ID":"872aa43a-f987-408d-a4c8-048df99f2244","Type":"ContainerStarted","Data":"48e7503ceaa99f6981ab16272dc72b253fd30aa5001497357508508a6ffd358b"} Dec 17 08:34:38 crc kubenswrapper[4966]: I1217 08:34:38.626480 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rzhjf" event={"ID":"ac78d231-7df6-4e74-aad8-166ba49f6176","Type":"ContainerStarted","Data":"017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b"} Dec 17 08:34:38 crc kubenswrapper[4966]: I1217 08:34:38.626610 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-rzhjf" podUID="ac78d231-7df6-4e74-aad8-166ba49f6176" containerName="registry-server" containerID="cri-o://017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b" gracePeriod=2 Dec 17 08:34:38 crc kubenswrapper[4966]: I1217 08:34:38.655826 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5tcwp" podStartSLOduration=2.598246851 podStartE2EDuration="2.655804282s" podCreationTimestamp="2025-12-17 08:34:36 +0000 UTC" firstStartedPulling="2025-12-17 08:34:37.918922686 +0000 UTC m=+813.463992628" lastFinishedPulling="2025-12-17 08:34:37.976480117 +0000 UTC m=+813.521550059" observedRunningTime="2025-12-17 08:34:38.652765837 +0000 UTC m=+814.197835819" watchObservedRunningTime="2025-12-17 08:34:38.655804282 +0000 UTC m=+814.200874244" Dec 17 08:34:38 crc kubenswrapper[4966]: I1217 08:34:38.979387 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rzhjf" Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.073420 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8rpv\" (UniqueName: \"kubernetes.io/projected/ac78d231-7df6-4e74-aad8-166ba49f6176-kube-api-access-t8rpv\") pod \"ac78d231-7df6-4e74-aad8-166ba49f6176\" (UID: \"ac78d231-7df6-4e74-aad8-166ba49f6176\") " Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.080079 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac78d231-7df6-4e74-aad8-166ba49f6176-kube-api-access-t8rpv" (OuterVolumeSpecName: "kube-api-access-t8rpv") pod "ac78d231-7df6-4e74-aad8-166ba49f6176" (UID: "ac78d231-7df6-4e74-aad8-166ba49f6176"). InnerVolumeSpecName "kube-api-access-t8rpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.175385 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8rpv\" (UniqueName: \"kubernetes.io/projected/ac78d231-7df6-4e74-aad8-166ba49f6176-kube-api-access-t8rpv\") on node \"crc\" DevicePath \"\"" Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.635926 4966 generic.go:334] "Generic (PLEG): container finished" podID="ac78d231-7df6-4e74-aad8-166ba49f6176" containerID="017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b" exitCode=0 Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.635998 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rzhjf" Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.636022 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rzhjf" event={"ID":"ac78d231-7df6-4e74-aad8-166ba49f6176","Type":"ContainerDied","Data":"017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b"} Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.636062 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rzhjf" event={"ID":"ac78d231-7df6-4e74-aad8-166ba49f6176","Type":"ContainerDied","Data":"1b0a3597564b7d5eeca8e9ca0dfc9b0a31d35a40a8efb6943529a4634d87540c"} Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.636078 4966 scope.go:117] "RemoveContainer" containerID="017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b" Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.660237 4966 scope.go:117] "RemoveContainer" containerID="017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b" Dec 17 08:34:39 crc kubenswrapper[4966]: E1217 08:34:39.660644 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b\": container with ID starting with 017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b not found: ID does not exist" containerID="017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b" Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.660721 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b"} err="failed to get container status \"017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b\": rpc error: code = NotFound desc = could not find container \"017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b\": container with ID starting with 017b539299212af3245ac0053236b26a40ab41965e61ea103e217d46e0e1883b not found: ID does not exist" Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.675896 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rzhjf"] Dec 17 08:34:39 crc kubenswrapper[4966]: I1217 08:34:39.681200 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-rzhjf"] Dec 17 08:34:40 crc kubenswrapper[4966]: I1217 08:34:40.839631 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac78d231-7df6-4e74-aad8-166ba49f6176" path="/var/lib/kubelet/pods/ac78d231-7df6-4e74-aad8-166ba49f6176/volumes" Dec 17 08:34:46 crc kubenswrapper[4966]: I1217 08:34:46.353702 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-5tcwp" Dec 17 08:34:46 crc kubenswrapper[4966]: I1217 08:34:46.354249 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5tcwp" Dec 17 08:34:46 crc kubenswrapper[4966]: I1217 08:34:46.371443 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-gvbzk" Dec 17 08:34:46 crc kubenswrapper[4966]: I1217 08:34:46.397449 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-5tcwp" Dec 17 08:34:46 crc kubenswrapper[4966]: I1217 08:34:46.729738 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-5tcwp" Dec 17 08:34:46 crc kubenswrapper[4966]: I1217 08:34:46.808061 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:34:46 crc kubenswrapper[4966]: I1217 08:34:46.808328 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:34:47 crc kubenswrapper[4966]: I1217 08:34:47.886123 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h"] Dec 17 08:34:47 crc kubenswrapper[4966]: E1217 08:34:47.886609 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac78d231-7df6-4e74-aad8-166ba49f6176" containerName="registry-server" Dec 17 08:34:47 crc kubenswrapper[4966]: I1217 08:34:47.886637 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac78d231-7df6-4e74-aad8-166ba49f6176" containerName="registry-server" Dec 17 08:34:47 crc kubenswrapper[4966]: I1217 08:34:47.886883 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac78d231-7df6-4e74-aad8-166ba49f6176" containerName="registry-server" Dec 17 08:34:47 crc kubenswrapper[4966]: I1217 08:34:47.889176 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:47 crc kubenswrapper[4966]: I1217 08:34:47.891602 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8297j" Dec 17 08:34:47 crc kubenswrapper[4966]: I1217 08:34:47.910029 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h"] Dec 17 08:34:47 crc kubenswrapper[4966]: I1217 08:34:47.999769 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-util\") pod \"332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:47.999809 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-bundle\") pod \"332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:47.999927 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l2nz\" (UniqueName: \"kubernetes.io/projected/8faaf4f3-ab4e-45b0-a494-2aac40eed118-kube-api-access-2l2nz\") pod \"332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:48.100852 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l2nz\" (UniqueName: \"kubernetes.io/projected/8faaf4f3-ab4e-45b0-a494-2aac40eed118-kube-api-access-2l2nz\") pod \"332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:48.101078 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-util\") pod \"332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:48.101119 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-bundle\") pod \"332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:48.102228 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-bundle\") pod \"332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:48.102411 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-util\") pod \"332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:48.122909 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l2nz\" (UniqueName: \"kubernetes.io/projected/8faaf4f3-ab4e-45b0-a494-2aac40eed118-kube-api-access-2l2nz\") pod \"332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:48.253649 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:48.666016 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h"] Dec 17 08:34:48 crc kubenswrapper[4966]: I1217 08:34:48.716013 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" event={"ID":"8faaf4f3-ab4e-45b0-a494-2aac40eed118","Type":"ContainerStarted","Data":"051794d243e18da94788abeac24137061bee4e20135b2ada375373aec4cabfc4"} Dec 17 08:34:49 crc kubenswrapper[4966]: I1217 08:34:49.728786 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" event={"ID":"8faaf4f3-ab4e-45b0-a494-2aac40eed118","Type":"ContainerDied","Data":"c42a3ac3874e181015588fba04519e49cf879513e15f162c79235dd22829d151"} Dec 17 08:34:49 crc kubenswrapper[4966]: I1217 08:34:49.728712 4966 generic.go:334] "Generic (PLEG): container finished" podID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerID="c42a3ac3874e181015588fba04519e49cf879513e15f162c79235dd22829d151" exitCode=0 Dec 17 08:34:50 crc kubenswrapper[4966]: I1217 08:34:50.739770 4966 generic.go:334] "Generic (PLEG): container finished" podID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerID="58ce9bc2808fbe8a74b3667334021973c8f4d77f33a264c41f3b5c0d0ac63d2b" exitCode=0 Dec 17 08:34:50 crc kubenswrapper[4966]: I1217 08:34:50.740818 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" event={"ID":"8faaf4f3-ab4e-45b0-a494-2aac40eed118","Type":"ContainerDied","Data":"58ce9bc2808fbe8a74b3667334021973c8f4d77f33a264c41f3b5c0d0ac63d2b"} Dec 17 08:34:51 crc kubenswrapper[4966]: I1217 08:34:51.747089 4966 generic.go:334] "Generic (PLEG): container finished" podID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerID="10ac27bf9d99db7e4dcfc241372744253b65bbea0c34e756845e809e3bcee18d" exitCode=0 Dec 17 08:34:51 crc kubenswrapper[4966]: I1217 08:34:51.747147 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" event={"ID":"8faaf4f3-ab4e-45b0-a494-2aac40eed118","Type":"ContainerDied","Data":"10ac27bf9d99db7e4dcfc241372744253b65bbea0c34e756845e809e3bcee18d"} Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.113017 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.280165 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-util\") pod \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.280315 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-bundle\") pod \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.280346 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l2nz\" (UniqueName: \"kubernetes.io/projected/8faaf4f3-ab4e-45b0-a494-2aac40eed118-kube-api-access-2l2nz\") pod \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\" (UID: \"8faaf4f3-ab4e-45b0-a494-2aac40eed118\") " Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.281079 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-bundle" (OuterVolumeSpecName: "bundle") pod "8faaf4f3-ab4e-45b0-a494-2aac40eed118" (UID: "8faaf4f3-ab4e-45b0-a494-2aac40eed118"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.296901 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8faaf4f3-ab4e-45b0-a494-2aac40eed118-kube-api-access-2l2nz" (OuterVolumeSpecName: "kube-api-access-2l2nz") pod "8faaf4f3-ab4e-45b0-a494-2aac40eed118" (UID: "8faaf4f3-ab4e-45b0-a494-2aac40eed118"). InnerVolumeSpecName "kube-api-access-2l2nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.299286 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-util" (OuterVolumeSpecName: "util") pod "8faaf4f3-ab4e-45b0-a494-2aac40eed118" (UID: "8faaf4f3-ab4e-45b0-a494-2aac40eed118"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.381943 4966 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-util\") on node \"crc\" DevicePath \"\"" Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.381982 4966 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8faaf4f3-ab4e-45b0-a494-2aac40eed118-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.381995 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l2nz\" (UniqueName: \"kubernetes.io/projected/8faaf4f3-ab4e-45b0-a494-2aac40eed118-kube-api-access-2l2nz\") on node \"crc\" DevicePath \"\"" Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.770217 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" event={"ID":"8faaf4f3-ab4e-45b0-a494-2aac40eed118","Type":"ContainerDied","Data":"051794d243e18da94788abeac24137061bee4e20135b2ada375373aec4cabfc4"} Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.770818 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="051794d243e18da94788abeac24137061bee4e20135b2ada375373aec4cabfc4" Dec 17 08:34:53 crc kubenswrapper[4966]: I1217 08:34:53.770446 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/332599f00c470fbfea7302cf68f0d07537a2d792b5c85cc6bac112d333k9c6h" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.508723 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv"] Dec 17 08:35:00 crc kubenswrapper[4966]: E1217 08:35:00.509537 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerName="pull" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.509553 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerName="pull" Dec 17 08:35:00 crc kubenswrapper[4966]: E1217 08:35:00.509567 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerName="extract" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.509576 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerName="extract" Dec 17 08:35:00 crc kubenswrapper[4966]: E1217 08:35:00.509585 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerName="util" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.509593 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerName="util" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.509727 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="8faaf4f3-ab4e-45b0-a494-2aac40eed118" containerName="extract" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.510239 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.514666 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-xv87h" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.568362 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv"] Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.687965 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf9nf\" (UniqueName: \"kubernetes.io/projected/98739565-9578-439e-b800-46d48c04546f-kube-api-access-wf9nf\") pod \"openstack-operator-controller-operator-6fc59d49b4-npkrv\" (UID: \"98739565-9578-439e-b800-46d48c04546f\") " pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.789118 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf9nf\" (UniqueName: \"kubernetes.io/projected/98739565-9578-439e-b800-46d48c04546f-kube-api-access-wf9nf\") pod \"openstack-operator-controller-operator-6fc59d49b4-npkrv\" (UID: \"98739565-9578-439e-b800-46d48c04546f\") " pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.812795 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf9nf\" (UniqueName: \"kubernetes.io/projected/98739565-9578-439e-b800-46d48c04546f-kube-api-access-wf9nf\") pod \"openstack-operator-controller-operator-6fc59d49b4-npkrv\" (UID: \"98739565-9578-439e-b800-46d48c04546f\") " pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" Dec 17 08:35:00 crc kubenswrapper[4966]: I1217 08:35:00.836074 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" Dec 17 08:35:01 crc kubenswrapper[4966]: I1217 08:35:01.123380 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv"] Dec 17 08:35:01 crc kubenswrapper[4966]: I1217 08:35:01.868122 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" event={"ID":"98739565-9578-439e-b800-46d48c04546f","Type":"ContainerStarted","Data":"c17138d58bff1cda88c98f32caba3f3537594a49c170137b2fcb39b0c4a3aeac"} Dec 17 08:35:08 crc kubenswrapper[4966]: I1217 08:35:08.925661 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" event={"ID":"98739565-9578-439e-b800-46d48c04546f","Type":"ContainerStarted","Data":"3dcf6aa321a41d4a9312a7f330c019d726899de77dd211fe2201a9cdcaf1a8e8"} Dec 17 08:35:08 crc kubenswrapper[4966]: I1217 08:35:08.926202 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" Dec 17 08:35:16 crc kubenswrapper[4966]: I1217 08:35:16.807744 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:35:16 crc kubenswrapper[4966]: I1217 08:35:16.808316 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:35:16 crc kubenswrapper[4966]: I1217 08:35:16.808363 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:35:16 crc kubenswrapper[4966]: I1217 08:35:16.808988 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6002081a3e53c22c985a84be589971720349cfd91c47bfefcfa37db148bb69f7"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 08:35:16 crc kubenswrapper[4966]: I1217 08:35:16.809044 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://6002081a3e53c22c985a84be589971720349cfd91c47bfefcfa37db148bb69f7" gracePeriod=600 Dec 17 08:35:17 crc kubenswrapper[4966]: I1217 08:35:17.007248 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="6002081a3e53c22c985a84be589971720349cfd91c47bfefcfa37db148bb69f7" exitCode=0 Dec 17 08:35:17 crc kubenswrapper[4966]: I1217 08:35:17.007324 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"6002081a3e53c22c985a84be589971720349cfd91c47bfefcfa37db148bb69f7"} Dec 17 08:35:17 crc kubenswrapper[4966]: I1217 08:35:17.007460 4966 scope.go:117] "RemoveContainer" containerID="ee72e03cd88492b874cbcd116164fd7756b1677dac83cef997b128cc487a5628" Dec 17 08:35:18 crc kubenswrapper[4966]: I1217 08:35:18.015343 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"dfe84a49239318b4c24aec015fc04c7e7ced77b13085152134e648f1545c4edf"} Dec 17 08:35:18 crc kubenswrapper[4966]: I1217 08:35:18.034934 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" podStartSLOduration=10.951898699000001 podStartE2EDuration="18.034900456s" podCreationTimestamp="2025-12-17 08:35:00 +0000 UTC" firstStartedPulling="2025-12-17 08:35:01.138048058 +0000 UTC m=+836.683118000" lastFinishedPulling="2025-12-17 08:35:08.221049815 +0000 UTC m=+843.766119757" observedRunningTime="2025-12-17 08:35:08.97931718 +0000 UTC m=+844.524387152" watchObservedRunningTime="2025-12-17 08:35:18.034900456 +0000 UTC m=+853.579970438" Dec 17 08:35:20 crc kubenswrapper[4966]: I1217 08:35:20.841068 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6fc59d49b4-npkrv" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.439319 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.440391 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.448135 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-mx6mr" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.462795 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-95949466-p69c8"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.464043 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.471904 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-ch6wz" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.479096 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.498939 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-95949466-p69c8"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.509813 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvjjx\" (UniqueName: \"kubernetes.io/projected/2eb3f60e-624a-43e1-a27a-48d466689c78-kube-api-access-jvjjx\") pod \"cinder-operator-controller-manager-5f98b4754f-dp5bm\" (UID: \"2eb3f60e-624a-43e1-a27a-48d466689c78\") " pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.533838 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.534692 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.536861 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-hnm22" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.553506 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.554533 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.559381 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-5jcks" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.576407 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.593188 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.601511 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.602024 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.606917 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-5l8tr" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.619134 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9prr2\" (UniqueName: \"kubernetes.io/projected/202f0a3b-a9da-4ce9-9297-10cba922d8ab-kube-api-access-9prr2\") pod \"barbican-operator-controller-manager-95949466-p69c8\" (UID: \"202f0a3b-a9da-4ce9-9297-10cba922d8ab\") " pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.619416 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjmkq\" (UniqueName: \"kubernetes.io/projected/1916840c-817c-4ad4-849f-4eb0f4787a79-kube-api-access-cjmkq\") pod \"glance-operator-controller-manager-767f9d7567-g7zdx\" (UID: \"1916840c-817c-4ad4-849f-4eb0f4787a79\") " pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.619533 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvjjx\" (UniqueName: \"kubernetes.io/projected/2eb3f60e-624a-43e1-a27a-48d466689c78-kube-api-access-jvjjx\") pod \"cinder-operator-controller-manager-5f98b4754f-dp5bm\" (UID: \"2eb3f60e-624a-43e1-a27a-48d466689c78\") " pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.619638 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4gct\" (UniqueName: \"kubernetes.io/projected/85b3221d-40ef-454b-920f-f0934adb7af1-kube-api-access-c4gct\") pod \"designate-operator-controller-manager-66f8b87655-l6ch2\" (UID: \"85b3221d-40ef-454b-920f-f0934adb7af1\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.623085 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.624102 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.627224 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-hw4wq" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.667934 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.682564 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.694798 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-84b495f78-tgd62"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.695628 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.698759 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.703614 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvjjx\" (UniqueName: \"kubernetes.io/projected/2eb3f60e-624a-43e1-a27a-48d466689c78-kube-api-access-jvjjx\") pod \"cinder-operator-controller-manager-5f98b4754f-dp5bm\" (UID: \"2eb3f60e-624a-43e1-a27a-48d466689c78\") " pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.704116 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-8l8k8" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.720858 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9prr2\" (UniqueName: \"kubernetes.io/projected/202f0a3b-a9da-4ce9-9297-10cba922d8ab-kube-api-access-9prr2\") pod \"barbican-operator-controller-manager-95949466-p69c8\" (UID: \"202f0a3b-a9da-4ce9-9297-10cba922d8ab\") " pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.720914 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjmkq\" (UniqueName: \"kubernetes.io/projected/1916840c-817c-4ad4-849f-4eb0f4787a79-kube-api-access-cjmkq\") pod \"glance-operator-controller-manager-767f9d7567-g7zdx\" (UID: \"1916840c-817c-4ad4-849f-4eb0f4787a79\") " pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.720954 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rbgc\" (UniqueName: \"kubernetes.io/projected/2803b782-fe84-4912-b7e6-528240fe14cb-kube-api-access-7rbgc\") pod \"heat-operator-controller-manager-59b8dcb766-w6wjs\" (UID: \"2803b782-fe84-4912-b7e6-528240fe14cb\") " pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.720975 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4gct\" (UniqueName: \"kubernetes.io/projected/85b3221d-40ef-454b-920f-f0934adb7af1-kube-api-access-c4gct\") pod \"designate-operator-controller-manager-66f8b87655-l6ch2\" (UID: \"85b3221d-40ef-454b-920f-f0934adb7af1\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.721012 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhd6g\" (UniqueName: \"kubernetes.io/projected/b288f5ca-8391-4157-8f02-886c88887df1-kube-api-access-rhd6g\") pod \"horizon-operator-controller-manager-6ccf486b9-64x68\" (UID: \"b288f5ca-8391-4157-8f02-886c88887df1\") " pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.728941 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.730214 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.746006 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-kqtdg" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.750182 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84b495f78-tgd62"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.755486 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4gct\" (UniqueName: \"kubernetes.io/projected/85b3221d-40ef-454b-920f-f0934adb7af1-kube-api-access-c4gct\") pod \"designate-operator-controller-manager-66f8b87655-l6ch2\" (UID: \"85b3221d-40ef-454b-920f-f0934adb7af1\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.769693 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.770517 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjmkq\" (UniqueName: \"kubernetes.io/projected/1916840c-817c-4ad4-849f-4eb0f4787a79-kube-api-access-cjmkq\") pod \"glance-operator-controller-manager-767f9d7567-g7zdx\" (UID: \"1916840c-817c-4ad4-849f-4eb0f4787a79\") " pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.772689 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9prr2\" (UniqueName: \"kubernetes.io/projected/202f0a3b-a9da-4ce9-9297-10cba922d8ab-kube-api-access-9prr2\") pod \"barbican-operator-controller-manager-95949466-p69c8\" (UID: \"202f0a3b-a9da-4ce9-9297-10cba922d8ab\") " pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.785018 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.785859 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.795184 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.795242 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-q4q7w" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.799034 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.799953 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.809772 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.812712 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.812862 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6522f" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.822276 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhd6g\" (UniqueName: \"kubernetes.io/projected/b288f5ca-8391-4157-8f02-886c88887df1-kube-api-access-rhd6g\") pod \"horizon-operator-controller-manager-6ccf486b9-64x68\" (UID: \"b288f5ca-8391-4157-8f02-886c88887df1\") " pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.822425 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz2gn\" (UniqueName: \"kubernetes.io/projected/9e86820d-0103-4900-b92a-ef86d1292689-kube-api-access-rz2gn\") pod \"ironic-operator-controller-manager-f458558d7-f98m5\" (UID: \"9e86820d-0103-4900-b92a-ef86d1292689\") " pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.822605 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.822697 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rbgc\" (UniqueName: \"kubernetes.io/projected/2803b782-fe84-4912-b7e6-528240fe14cb-kube-api-access-7rbgc\") pod \"heat-operator-controller-manager-59b8dcb766-w6wjs\" (UID: \"2803b782-fe84-4912-b7e6-528240fe14cb\") " pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.822787 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4f2k\" (UniqueName: \"kubernetes.io/projected/0be98506-68e8-438f-a46b-2225dd61b871-kube-api-access-w4f2k\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.850606 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhd6g\" (UniqueName: \"kubernetes.io/projected/b288f5ca-8391-4157-8f02-886c88887df1-kube-api-access-rhd6g\") pod \"horizon-operator-controller-manager-6ccf486b9-64x68\" (UID: \"b288f5ca-8391-4157-8f02-886c88887df1\") " pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.854404 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.860092 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.860924 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.865201 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-9ksh4" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.874341 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.879219 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.885277 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.895435 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rbgc\" (UniqueName: \"kubernetes.io/projected/2803b782-fe84-4912-b7e6-528240fe14cb-kube-api-access-7rbgc\") pod \"heat-operator-controller-manager-59b8dcb766-w6wjs\" (UID: \"2803b782-fe84-4912-b7e6-528240fe14cb\") " pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.923561 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hggzx\" (UniqueName: \"kubernetes.io/projected/8157370d-5a0a-4d55-98be-1061654d3a37-kube-api-access-hggzx\") pod \"mariadb-operator-controller-manager-f76f4954c-mfghn\" (UID: \"8157370d-5a0a-4d55-98be-1061654d3a37\") " pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.923618 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.923677 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4f2k\" (UniqueName: \"kubernetes.io/projected/0be98506-68e8-438f-a46b-2225dd61b871-kube-api-access-w4f2k\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.923711 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz2gn\" (UniqueName: \"kubernetes.io/projected/9e86820d-0103-4900-b92a-ef86d1292689-kube-api-access-rz2gn\") pod \"ironic-operator-controller-manager-f458558d7-f98m5\" (UID: \"9e86820d-0103-4900-b92a-ef86d1292689\") " pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.923744 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmtvz\" (UniqueName: \"kubernetes.io/projected/c1229400-ebf1-4fb6-a8a0-4f7292c22dfc-kube-api-access-hmtvz\") pod \"keystone-operator-controller-manager-5c7cbf548f-zmxmt\" (UID: \"c1229400-ebf1-4fb6-a8a0-4f7292c22dfc\") " pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.923760 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fcqj\" (UniqueName: \"kubernetes.io/projected/d412de43-e6c6-4c8a-a947-3811cad9d6a8-kube-api-access-6fcqj\") pod \"manila-operator-controller-manager-5fdd9786f7-tv6n9\" (UID: \"d412de43-e6c6-4c8a-a947-3811cad9d6a8\") " pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" Dec 17 08:35:38 crc kubenswrapper[4966]: E1217 08:35:38.924098 4966 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:38 crc kubenswrapper[4966]: E1217 08:35:38.924143 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert podName:0be98506-68e8-438f-a46b-2225dd61b871 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:39.424127375 +0000 UTC m=+874.969197317 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert") pod "infra-operator-controller-manager-84b495f78-tgd62" (UID: "0be98506-68e8-438f-a46b-2225dd61b871") : secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.934155 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.935504 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.939855 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-wrs82" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.953402 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.959236 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.960374 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.968024 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-nwskb" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.971449 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.972515 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.976515 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz2gn\" (UniqueName: \"kubernetes.io/projected/9e86820d-0103-4900-b92a-ef86d1292689-kube-api-access-rz2gn\") pod \"ironic-operator-controller-manager-f458558d7-f98m5\" (UID: \"9e86820d-0103-4900-b92a-ef86d1292689\") " pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.978655 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-wz9ll" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.981544 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr"] Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.983173 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4f2k\" (UniqueName: \"kubernetes.io/projected/0be98506-68e8-438f-a46b-2225dd61b871-kube-api-access-w4f2k\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:38 crc kubenswrapper[4966]: I1217 08:35:38.989020 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.082298 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g2qv\" (UniqueName: \"kubernetes.io/projected/ae9e23d6-8355-49f3-8558-1dcc50b0d12e-kube-api-access-7g2qv\") pod \"nova-operator-controller-manager-5fbbf8b6cc-6z6kr\" (UID: \"ae9e23d6-8355-49f3-8558-1dcc50b0d12e\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.088468 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmtvz\" (UniqueName: \"kubernetes.io/projected/c1229400-ebf1-4fb6-a8a0-4f7292c22dfc-kube-api-access-hmtvz\") pod \"keystone-operator-controller-manager-5c7cbf548f-zmxmt\" (UID: \"c1229400-ebf1-4fb6-a8a0-4f7292c22dfc\") " pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.088699 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fcqj\" (UniqueName: \"kubernetes.io/projected/d412de43-e6c6-4c8a-a947-3811cad9d6a8-kube-api-access-6fcqj\") pod \"manila-operator-controller-manager-5fdd9786f7-tv6n9\" (UID: \"d412de43-e6c6-4c8a-a947-3811cad9d6a8\") " pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.088846 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brngw\" (UniqueName: \"kubernetes.io/projected/62fbcfd8-be0b-4b98-bc08-337173ee2758-kube-api-access-brngw\") pod \"octavia-operator-controller-manager-68c649d9d-cr8zh\" (UID: \"62fbcfd8-be0b-4b98-bc08-337173ee2758\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.097923 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.097986 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hggzx\" (UniqueName: \"kubernetes.io/projected/8157370d-5a0a-4d55-98be-1061654d3a37-kube-api-access-hggzx\") pod \"mariadb-operator-controller-manager-f76f4954c-mfghn\" (UID: \"8157370d-5a0a-4d55-98be-1061654d3a37\") " pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.098164 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfmlg\" (UniqueName: \"kubernetes.io/projected/9c79b45a-796b-4bba-9daa-80d2c7e7d55b-kube-api-access-gfmlg\") pod \"neutron-operator-controller-manager-7cd87b778f-b2q7b\" (UID: \"9c79b45a-796b-4bba-9daa-80d2c7e7d55b\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.191859 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmtvz\" (UniqueName: \"kubernetes.io/projected/c1229400-ebf1-4fb6-a8a0-4f7292c22dfc-kube-api-access-hmtvz\") pod \"keystone-operator-controller-manager-5c7cbf548f-zmxmt\" (UID: \"c1229400-ebf1-4fb6-a8a0-4f7292c22dfc\") " pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.196323 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fcqj\" (UniqueName: \"kubernetes.io/projected/d412de43-e6c6-4c8a-a947-3811cad9d6a8-kube-api-access-6fcqj\") pod \"manila-operator-controller-manager-5fdd9786f7-tv6n9\" (UID: \"d412de43-e6c6-4c8a-a947-3811cad9d6a8\") " pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.201710 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfmlg\" (UniqueName: \"kubernetes.io/projected/9c79b45a-796b-4bba-9daa-80d2c7e7d55b-kube-api-access-gfmlg\") pod \"neutron-operator-controller-manager-7cd87b778f-b2q7b\" (UID: \"9c79b45a-796b-4bba-9daa-80d2c7e7d55b\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.201802 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g2qv\" (UniqueName: \"kubernetes.io/projected/ae9e23d6-8355-49f3-8558-1dcc50b0d12e-kube-api-access-7g2qv\") pod \"nova-operator-controller-manager-5fbbf8b6cc-6z6kr\" (UID: \"ae9e23d6-8355-49f3-8558-1dcc50b0d12e\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.201853 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brngw\" (UniqueName: \"kubernetes.io/projected/62fbcfd8-be0b-4b98-bc08-337173ee2758-kube-api-access-brngw\") pod \"octavia-operator-controller-manager-68c649d9d-cr8zh\" (UID: \"62fbcfd8-be0b-4b98-bc08-337173ee2758\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.202975 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hggzx\" (UniqueName: \"kubernetes.io/projected/8157370d-5a0a-4d55-98be-1061654d3a37-kube-api-access-hggzx\") pod \"mariadb-operator-controller-manager-f76f4954c-mfghn\" (UID: \"8157370d-5a0a-4d55-98be-1061654d3a37\") " pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.231055 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.242256 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.246589 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g2qv\" (UniqueName: \"kubernetes.io/projected/ae9e23d6-8355-49f3-8558-1dcc50b0d12e-kube-api-access-7g2qv\") pod \"nova-operator-controller-manager-5fbbf8b6cc-6z6kr\" (UID: \"ae9e23d6-8355-49f3-8558-1dcc50b0d12e\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.248041 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfmlg\" (UniqueName: \"kubernetes.io/projected/9c79b45a-796b-4bba-9daa-80d2c7e7d55b-kube-api-access-gfmlg\") pod \"neutron-operator-controller-manager-7cd87b778f-b2q7b\" (UID: \"9c79b45a-796b-4bba-9daa-80d2c7e7d55b\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.266809 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brngw\" (UniqueName: \"kubernetes.io/projected/62fbcfd8-be0b-4b98-bc08-337173ee2758-kube-api-access-brngw\") pod \"octavia-operator-controller-manager-68c649d9d-cr8zh\" (UID: \"62fbcfd8-be0b-4b98-bc08-337173ee2758\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.268277 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.276467 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.277692 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.287583 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.288266 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-kxhv9" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.294562 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.300567 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.320574 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.323027 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.329240 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-82dtb" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.337661 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.338804 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.340610 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.354115 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.369504 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-l6dgw" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.371232 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.372333 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.376633 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.377760 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.378076 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.378921 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-88p5r" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.382325 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-sdg7g" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.387102 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.388122 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.391414 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.400353 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.400950 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.409852 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-h2qsh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.412399 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.413624 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ld29\" (UniqueName: \"kubernetes.io/projected/cfe05261-97ae-4fd8-b178-1a0679c689da-kube-api-access-5ld29\") pod \"ovn-operator-controller-manager-bf6d4f946-fj2v2\" (UID: \"cfe05261-97ae-4fd8-b178-1a0679c689da\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.413687 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl5f6\" (UniqueName: \"kubernetes.io/projected/add4eb97-35a5-4070-ac9f-e14eb5741aac-kube-api-access-wl5f6\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.413727 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.416932 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79vbf\" (UniqueName: \"kubernetes.io/projected/dd104116-5bb1-421b-ae58-9b7b6e6f79e5-kube-api-access-79vbf\") pod \"placement-operator-controller-manager-8665b56d78-mv4mc\" (UID: \"dd104116-5bb1-421b-ae58-9b7b6e6f79e5\") " pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.434118 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.435657 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.439341 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.447009 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.448068 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.455662 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.456154 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.456281 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-m64k2" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.456429 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-9dkpg" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.456958 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.490543 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.491746 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.498244 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.499009 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.505863 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-wb4kx" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517688 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517756 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq5qt\" (UniqueName: \"kubernetes.io/projected/64c2ea0f-3e8c-4872-b130-df049355e147-kube-api-access-wq5qt\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517782 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517806 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79vbf\" (UniqueName: \"kubernetes.io/projected/dd104116-5bb1-421b-ae58-9b7b6e6f79e5-kube-api-access-79vbf\") pod \"placement-operator-controller-manager-8665b56d78-mv4mc\" (UID: \"dd104116-5bb1-421b-ae58-9b7b6e6f79e5\") " pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517828 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ld29\" (UniqueName: \"kubernetes.io/projected/cfe05261-97ae-4fd8-b178-1a0679c689da-kube-api-access-5ld29\") pod \"ovn-operator-controller-manager-bf6d4f946-fj2v2\" (UID: \"cfe05261-97ae-4fd8-b178-1a0679c689da\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517856 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x9p7\" (UniqueName: \"kubernetes.io/projected/861f02ab-0025-4dbb-b020-e02561b3830e-kube-api-access-6x9p7\") pod \"watcher-operator-controller-manager-55f78b7c4c-z9pzc\" (UID: \"861f02ab-0025-4dbb-b020-e02561b3830e\") " pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517907 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl5f6\" (UniqueName: \"kubernetes.io/projected/add4eb97-35a5-4070-ac9f-e14eb5741aac-kube-api-access-wl5f6\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517928 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqj2s\" (UniqueName: \"kubernetes.io/projected/1458fa23-50b0-4dad-bed4-d9859806d337-kube-api-access-wqj2s\") pod \"telemetry-operator-controller-manager-97d456b9-g6xkf\" (UID: \"1458fa23-50b0-4dad-bed4-d9859806d337\") " pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517962 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbtjh\" (UniqueName: \"kubernetes.io/projected/2cc894eb-b898-4c2e-b379-7a48d08d6293-kube-api-access-lbtjh\") pod \"swift-operator-controller-manager-5c6df8f9-rcsdh\" (UID: \"2cc894eb-b898-4c2e-b379-7a48d08d6293\") " pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.517980 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.518003 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvfvf\" (UniqueName: \"kubernetes.io/projected/5277f562-0c42-4ca8-b8a5-aa345d02fca7-kube-api-access-kvfvf\") pod \"test-operator-controller-manager-756ccf86c7-ksgt7\" (UID: \"5277f562-0c42-4ca8-b8a5-aa345d02fca7\") " pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.518040 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:39 crc kubenswrapper[4966]: E1217 08:35:39.518158 4966 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:39 crc kubenswrapper[4966]: E1217 08:35:39.518206 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert podName:0be98506-68e8-438f-a46b-2225dd61b871 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:40.518193021 +0000 UTC m=+876.063262963 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert") pod "infra-operator-controller-manager-84b495f78-tgd62" (UID: "0be98506-68e8-438f-a46b-2225dd61b871") : secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:39 crc kubenswrapper[4966]: E1217 08:35:39.518734 4966 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:39 crc kubenswrapper[4966]: E1217 08:35:39.518763 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert podName:add4eb97-35a5-4070-ac9f-e14eb5741aac nodeName:}" failed. No retries permitted until 2025-12-17 08:35:40.018755966 +0000 UTC m=+875.563825908 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" (UID: "add4eb97-35a5-4070-ac9f-e14eb5741aac") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.552827 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ld29\" (UniqueName: \"kubernetes.io/projected/cfe05261-97ae-4fd8-b178-1a0679c689da-kube-api-access-5ld29\") pod \"ovn-operator-controller-manager-bf6d4f946-fj2v2\" (UID: \"cfe05261-97ae-4fd8-b178-1a0679c689da\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.556585 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.562024 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl5f6\" (UniqueName: \"kubernetes.io/projected/add4eb97-35a5-4070-ac9f-e14eb5741aac-kube-api-access-wl5f6\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.570684 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79vbf\" (UniqueName: \"kubernetes.io/projected/dd104116-5bb1-421b-ae58-9b7b6e6f79e5-kube-api-access-79vbf\") pod \"placement-operator-controller-manager-8665b56d78-mv4mc\" (UID: \"dd104116-5bb1-421b-ae58-9b7b6e6f79e5\") " pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.622347 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq5qt\" (UniqueName: \"kubernetes.io/projected/64c2ea0f-3e8c-4872-b130-df049355e147-kube-api-access-wq5qt\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.622422 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.622472 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x9p7\" (UniqueName: \"kubernetes.io/projected/861f02ab-0025-4dbb-b020-e02561b3830e-kube-api-access-6x9p7\") pod \"watcher-operator-controller-manager-55f78b7c4c-z9pzc\" (UID: \"861f02ab-0025-4dbb-b020-e02561b3830e\") " pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.622510 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqj2s\" (UniqueName: \"kubernetes.io/projected/1458fa23-50b0-4dad-bed4-d9859806d337-kube-api-access-wqj2s\") pod \"telemetry-operator-controller-manager-97d456b9-g6xkf\" (UID: \"1458fa23-50b0-4dad-bed4-d9859806d337\") " pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.622539 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbtjh\" (UniqueName: \"kubernetes.io/projected/2cc894eb-b898-4c2e-b379-7a48d08d6293-kube-api-access-lbtjh\") pod \"swift-operator-controller-manager-5c6df8f9-rcsdh\" (UID: \"2cc894eb-b898-4c2e-b379-7a48d08d6293\") " pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.622574 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvfvf\" (UniqueName: \"kubernetes.io/projected/5277f562-0c42-4ca8-b8a5-aa345d02fca7-kube-api-access-kvfvf\") pod \"test-operator-controller-manager-756ccf86c7-ksgt7\" (UID: \"5277f562-0c42-4ca8-b8a5-aa345d02fca7\") " pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.622603 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dhkq\" (UniqueName: \"kubernetes.io/projected/f485453d-7629-41e7-a1c7-2c0b284bb551-kube-api-access-8dhkq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-8fvcq\" (UID: \"f485453d-7629-41e7-a1c7-2c0b284bb551\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.622657 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:39 crc kubenswrapper[4966]: E1217 08:35:39.622804 4966 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 17 08:35:39 crc kubenswrapper[4966]: E1217 08:35:39.622854 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:40.122834661 +0000 UTC m=+875.667904603 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "webhook-server-cert" not found Dec 17 08:35:39 crc kubenswrapper[4966]: E1217 08:35:39.624051 4966 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 17 08:35:39 crc kubenswrapper[4966]: E1217 08:35:39.624089 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:40.124075935 +0000 UTC m=+875.669145877 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "metrics-server-cert" not found Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.721577 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x9p7\" (UniqueName: \"kubernetes.io/projected/861f02ab-0025-4dbb-b020-e02561b3830e-kube-api-access-6x9p7\") pod \"watcher-operator-controller-manager-55f78b7c4c-z9pzc\" (UID: \"861f02ab-0025-4dbb-b020-e02561b3830e\") " pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.722496 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbtjh\" (UniqueName: \"kubernetes.io/projected/2cc894eb-b898-4c2e-b379-7a48d08d6293-kube-api-access-lbtjh\") pod \"swift-operator-controller-manager-5c6df8f9-rcsdh\" (UID: \"2cc894eb-b898-4c2e-b379-7a48d08d6293\") " pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.723600 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dhkq\" (UniqueName: \"kubernetes.io/projected/f485453d-7629-41e7-a1c7-2c0b284bb551-kube-api-access-8dhkq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-8fvcq\" (UID: \"f485453d-7629-41e7-a1c7-2c0b284bb551\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.723934 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.725415 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq5qt\" (UniqueName: \"kubernetes.io/projected/64c2ea0f-3e8c-4872-b130-df049355e147-kube-api-access-wq5qt\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.726575 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqj2s\" (UniqueName: \"kubernetes.io/projected/1458fa23-50b0-4dad-bed4-d9859806d337-kube-api-access-wqj2s\") pod \"telemetry-operator-controller-manager-97d456b9-g6xkf\" (UID: \"1458fa23-50b0-4dad-bed4-d9859806d337\") " pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.748642 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvfvf\" (UniqueName: \"kubernetes.io/projected/5277f562-0c42-4ca8-b8a5-aa345d02fca7-kube-api-access-kvfvf\") pod \"test-operator-controller-manager-756ccf86c7-ksgt7\" (UID: \"5277f562-0c42-4ca8-b8a5-aa345d02fca7\") " pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.768079 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.794124 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.796842 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dhkq\" (UniqueName: \"kubernetes.io/projected/f485453d-7629-41e7-a1c7-2c0b284bb551-kube-api-access-8dhkq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-8fvcq\" (UID: \"f485453d-7629-41e7-a1c7-2c0b284bb551\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.805416 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-95949466-p69c8"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.810820 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.855247 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.856497 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.953491 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm"] Dec 17 08:35:39 crc kubenswrapper[4966]: I1217 08:35:39.955293 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:39.999934 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx"] Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.054338 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:40 crc kubenswrapper[4966]: E1217 08:35:40.054496 4966 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:40 crc kubenswrapper[4966]: E1217 08:35:40.054560 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert podName:add4eb97-35a5-4070-ac9f-e14eb5741aac nodeName:}" failed. No retries permitted until 2025-12-17 08:35:41.054542399 +0000 UTC m=+876.599612341 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" (UID: "add4eb97-35a5-4070-ac9f-e14eb5741aac") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.157910 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.157970 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:40 crc kubenswrapper[4966]: E1217 08:35:40.158090 4966 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 17 08:35:40 crc kubenswrapper[4966]: E1217 08:35:40.158137 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:41.15812255 +0000 UTC m=+876.703192492 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "metrics-server-cert" not found Dec 17 08:35:40 crc kubenswrapper[4966]: E1217 08:35:40.158484 4966 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 17 08:35:40 crc kubenswrapper[4966]: E1217 08:35:40.158507 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:41.15850026 +0000 UTC m=+876.703570202 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "webhook-server-cert" not found Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.241594 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68"] Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.309028 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-795j5"] Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.310854 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.320932 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" event={"ID":"1916840c-817c-4ad4-849f-4eb0f4787a79","Type":"ContainerStarted","Data":"0d17ec474a2467d7ad88e3724dd20c900d8302678ac2c71358ec631209a4bd3d"} Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.343003 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" event={"ID":"2eb3f60e-624a-43e1-a27a-48d466689c78","Type":"ContainerStarted","Data":"c0eeb08af2893d5d73e92993257a3c7f729d3e5d0bc624520e7461ff2d4f1752"} Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.348723 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-795j5"] Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.359888 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-utilities\") pod \"certified-operators-795j5\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.359935 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-catalog-content\") pod \"certified-operators-795j5\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.359982 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn5d5\" (UniqueName: \"kubernetes.io/projected/057025a8-6d42-4a30-8156-b041e8c4d9aa-kube-api-access-zn5d5\") pod \"certified-operators-795j5\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.394101 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" event={"ID":"202f0a3b-a9da-4ce9-9297-10cba922d8ab","Type":"ContainerStarted","Data":"0e3aafd06625d2e09f313bab8e7e3108cd46be451916f3ea41e6bc2001f85c08"} Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.438990 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs"] Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.461989 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-utilities\") pod \"certified-operators-795j5\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.462031 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-catalog-content\") pod \"certified-operators-795j5\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.462077 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn5d5\" (UniqueName: \"kubernetes.io/projected/057025a8-6d42-4a30-8156-b041e8c4d9aa-kube-api-access-zn5d5\") pod \"certified-operators-795j5\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.462863 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-utilities\") pod \"certified-operators-795j5\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.463213 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-catalog-content\") pod \"certified-operators-795j5\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.465519 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2"] Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.496327 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn5d5\" (UniqueName: \"kubernetes.io/projected/057025a8-6d42-4a30-8156-b041e8c4d9aa-kube-api-access-zn5d5\") pod \"certified-operators-795j5\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: W1217 08:35:40.543797 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85b3221d_40ef_454b_920f_f0934adb7af1.slice/crio-dd134b2b8b92b738d78cf2cf19373ee647f36ed2181699914886a3043e739ca0 WatchSource:0}: Error finding container dd134b2b8b92b738d78cf2cf19373ee647f36ed2181699914886a3043e739ca0: Status 404 returned error can't find the container with id dd134b2b8b92b738d78cf2cf19373ee647f36ed2181699914886a3043e739ca0 Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.562955 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:40 crc kubenswrapper[4966]: E1217 08:35:40.563104 4966 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:40 crc kubenswrapper[4966]: E1217 08:35:40.563152 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert podName:0be98506-68e8-438f-a46b-2225dd61b871 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:42.563135154 +0000 UTC m=+878.108205096 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert") pod "infra-operator-controller-manager-84b495f78-tgd62" (UID: "0be98506-68e8-438f-a46b-2225dd61b871") : secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.679221 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.696937 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5"] Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.814595 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9"] Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.882945 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt"] Dec 17 08:35:40 crc kubenswrapper[4966]: W1217 08:35:40.890087 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd412de43_e6c6_4c8a_a947_3811cad9d6a8.slice/crio-4d922c80b2194eb3477b36f783d4862b219d482e73a0266927b3ba8ef75c8df9 WatchSource:0}: Error finding container 4d922c80b2194eb3477b36f783d4862b219d482e73a0266927b3ba8ef75c8df9: Status 404 returned error can't find the container with id 4d922c80b2194eb3477b36f783d4862b219d482e73a0266927b3ba8ef75c8df9 Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.946554 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn"] Dec 17 08:35:40 crc kubenswrapper[4966]: I1217 08:35:40.995436 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b"] Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.077058 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.077239 4966 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.077325 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert podName:add4eb97-35a5-4070-ac9f-e14eb5741aac nodeName:}" failed. No retries permitted until 2025-12-17 08:35:43.077288832 +0000 UTC m=+878.622358774 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" (UID: "add4eb97-35a5-4070-ac9f-e14eb5741aac") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.163980 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh"] Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.181852 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.181935 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.182107 4966 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.182131 4966 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.182178 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:43.182165817 +0000 UTC m=+878.727235759 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "metrics-server-cert" not found Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.182193 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:43.182187937 +0000 UTC m=+878.727257879 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "webhook-server-cert" not found Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.293006 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh"] Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.362189 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc"] Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.372302 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf"] Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.384929 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr"] Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.409701 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2"] Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.415646 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" event={"ID":"2803b782-fe84-4912-b7e6-528240fe14cb","Type":"ContainerStarted","Data":"c61bdfdedb3dabf1e0c76e426a8089d16ba611689ba027a5245c92f1c95188d3"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.439918 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc"] Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.443543 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kvfvf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-756ccf86c7-ksgt7_openstack-operators(5277f562-0c42-4ca8-b8a5-aa345d02fca7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.444788 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" podUID="5277f562-0c42-4ca8-b8a5-aa345d02fca7" Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.449670 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" event={"ID":"b288f5ca-8391-4157-8f02-886c88887df1","Type":"ContainerStarted","Data":"8205a2b6e9b48d14a8804079069942ef331f8fb2bafc69d19b16630e2511556c"} Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.450296 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6x9p7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-55f78b7c4c-z9pzc_openstack-operators(861f02ab-0025-4dbb-b020-e02561b3830e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.453735 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" podUID="861f02ab-0025-4dbb-b020-e02561b3830e" Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.459772 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" event={"ID":"d412de43-e6c6-4c8a-a947-3811cad9d6a8","Type":"ContainerStarted","Data":"4d922c80b2194eb3477b36f783d4862b219d482e73a0266927b3ba8ef75c8df9"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.463659 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" event={"ID":"ae9e23d6-8355-49f3-8558-1dcc50b0d12e","Type":"ContainerStarted","Data":"54d2dce7a2a30194e88ffec2f5094f0f9f6572f734f96c949f8583c7cf9a6578"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.473693 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" event={"ID":"62fbcfd8-be0b-4b98-bc08-337173ee2758","Type":"ContainerStarted","Data":"5f7f3b20a731666157ce86d85847c25d8f5fb0d90331c0bb45c9d05850750398"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.478555 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" event={"ID":"9c79b45a-796b-4bba-9daa-80d2c7e7d55b","Type":"ContainerStarted","Data":"eccffbc415faec8da161c47c078b952db0b8b5c19bc5a071051f4f33dd0cfe92"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.506146 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" event={"ID":"c1229400-ebf1-4fb6-a8a0-4f7292c22dfc","Type":"ContainerStarted","Data":"5df4d1eac9c3ddd20b19d9ee1f316815b6feacccab75bdc3d6165547b618f919"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.509392 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7"] Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.509861 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" event={"ID":"1458fa23-50b0-4dad-bed4-d9859806d337","Type":"ContainerStarted","Data":"250075f07a61aa30e1cd808cf066b15ed387001d2de7174d180556c985973ee2"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.527085 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" event={"ID":"8157370d-5a0a-4d55-98be-1061654d3a37","Type":"ContainerStarted","Data":"223a11940d4c8f5fd9c2ec377c3758d0482ee443376108ed9550dcc4d6859ef2"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.537067 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" event={"ID":"9e86820d-0103-4900-b92a-ef86d1292689","Type":"ContainerStarted","Data":"ec7dbe9048443a0a3261a22a93c8f1db19d91b2fc5765910589fde875a3bca2d"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.543150 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" event={"ID":"cfe05261-97ae-4fd8-b178-1a0679c689da","Type":"ContainerStarted","Data":"270863c1e5b93604fec5efb4228a8120d4559cad11b1da29ea8e8fc55b77041a"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.548863 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" event={"ID":"85b3221d-40ef-454b-920f-f0934adb7af1","Type":"ContainerStarted","Data":"dd134b2b8b92b738d78cf2cf19373ee647f36ed2181699914886a3043e739ca0"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.554414 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" event={"ID":"2cc894eb-b898-4c2e-b379-7a48d08d6293","Type":"ContainerStarted","Data":"e56389c33c4bc5718870d0ac2071eea34e1c3254cc0cf989a3d3659c143afcec"} Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.575242 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq"] Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.587538 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8dhkq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-8fvcq_openstack-operators(f485453d-7629-41e7-a1c7-2c0b284bb551): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 17 08:35:41 crc kubenswrapper[4966]: E1217 08:35:41.588789 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" podUID="f485453d-7629-41e7-a1c7-2c0b284bb551" Dec 17 08:35:41 crc kubenswrapper[4966]: I1217 08:35:41.659917 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-795j5"] Dec 17 08:35:42 crc kubenswrapper[4966]: I1217 08:35:42.578545 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" event={"ID":"861f02ab-0025-4dbb-b020-e02561b3830e","Type":"ContainerStarted","Data":"93ee42034e76065e31ad57f372a227719dc23aa5cbfc68f75eea02d5e170480f"} Dec 17 08:35:42 crc kubenswrapper[4966]: E1217 08:35:42.585173 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" podUID="861f02ab-0025-4dbb-b020-e02561b3830e" Dec 17 08:35:42 crc kubenswrapper[4966]: I1217 08:35:42.596271 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" event={"ID":"dd104116-5bb1-421b-ae58-9b7b6e6f79e5","Type":"ContainerStarted","Data":"8411bf8193713c6777a007fcab08a191a7e582c3cf9e5c786ed9bcbe7d7fb1fe"} Dec 17 08:35:42 crc kubenswrapper[4966]: I1217 08:35:42.598586 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" event={"ID":"5277f562-0c42-4ca8-b8a5-aa345d02fca7","Type":"ContainerStarted","Data":"52c297947dca29f55a4604f7a0fd894fff957bdacfb1be8116e02928daebb2a4"} Dec 17 08:35:42 crc kubenswrapper[4966]: E1217 08:35:42.600561 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" podUID="5277f562-0c42-4ca8-b8a5-aa345d02fca7" Dec 17 08:35:42 crc kubenswrapper[4966]: I1217 08:35:42.608288 4966 generic.go:334] "Generic (PLEG): container finished" podID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerID="0e28cef0856b356411a6187a94b70704373c7002299159e6be4c1ba2299eabb6" exitCode=0 Dec 17 08:35:42 crc kubenswrapper[4966]: I1217 08:35:42.608369 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795j5" event={"ID":"057025a8-6d42-4a30-8156-b041e8c4d9aa","Type":"ContainerDied","Data":"0e28cef0856b356411a6187a94b70704373c7002299159e6be4c1ba2299eabb6"} Dec 17 08:35:42 crc kubenswrapper[4966]: I1217 08:35:42.608389 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795j5" event={"ID":"057025a8-6d42-4a30-8156-b041e8c4d9aa","Type":"ContainerStarted","Data":"517b2299953334b07dce6bfebdc2344e8d8354ba2b67d15f9e75081fb0845778"} Dec 17 08:35:42 crc kubenswrapper[4966]: I1217 08:35:42.628359 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" event={"ID":"f485453d-7629-41e7-a1c7-2c0b284bb551","Type":"ContainerStarted","Data":"d8bf51205cea00ee967afc9ffe3894b10dc4ea2a1bf151fa35bc60f5f89f52fa"} Dec 17 08:35:42 crc kubenswrapper[4966]: E1217 08:35:42.629457 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" podUID="f485453d-7629-41e7-a1c7-2c0b284bb551" Dec 17 08:35:42 crc kubenswrapper[4966]: I1217 08:35:42.653214 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:42 crc kubenswrapper[4966]: E1217 08:35:42.653951 4966 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:42 crc kubenswrapper[4966]: E1217 08:35:42.653996 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert podName:0be98506-68e8-438f-a46b-2225dd61b871 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:46.653982956 +0000 UTC m=+882.199052898 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert") pod "infra-operator-controller-manager-84b495f78-tgd62" (UID: "0be98506-68e8-438f-a46b-2225dd61b871") : secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:43 crc kubenswrapper[4966]: I1217 08:35:43.159385 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:43 crc kubenswrapper[4966]: E1217 08:35:43.159528 4966 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:43 crc kubenswrapper[4966]: E1217 08:35:43.159580 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert podName:add4eb97-35a5-4070-ac9f-e14eb5741aac nodeName:}" failed. No retries permitted until 2025-12-17 08:35:47.159566688 +0000 UTC m=+882.704636630 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" (UID: "add4eb97-35a5-4070-ac9f-e14eb5741aac") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:43 crc kubenswrapper[4966]: I1217 08:35:43.261078 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:43 crc kubenswrapper[4966]: I1217 08:35:43.261144 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:43 crc kubenswrapper[4966]: E1217 08:35:43.261294 4966 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 17 08:35:43 crc kubenswrapper[4966]: E1217 08:35:43.261320 4966 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 17 08:35:43 crc kubenswrapper[4966]: E1217 08:35:43.261367 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:47.261353288 +0000 UTC m=+882.806423230 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "metrics-server-cert" not found Dec 17 08:35:43 crc kubenswrapper[4966]: E1217 08:35:43.261383 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:47.261375409 +0000 UTC m=+882.806445351 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "webhook-server-cert" not found Dec 17 08:35:43 crc kubenswrapper[4966]: E1217 08:35:43.663561 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" podUID="f485453d-7629-41e7-a1c7-2c0b284bb551" Dec 17 08:35:43 crc kubenswrapper[4966]: E1217 08:35:43.664355 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" podUID="861f02ab-0025-4dbb-b020-e02561b3830e" Dec 17 08:35:43 crc kubenswrapper[4966]: E1217 08:35:43.664817 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" podUID="5277f562-0c42-4ca8-b8a5-aa345d02fca7" Dec 17 08:35:46 crc kubenswrapper[4966]: I1217 08:35:46.717440 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:46 crc kubenswrapper[4966]: E1217 08:35:46.717693 4966 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:46 crc kubenswrapper[4966]: E1217 08:35:46.718185 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert podName:0be98506-68e8-438f-a46b-2225dd61b871 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:54.718161126 +0000 UTC m=+890.263231138 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert") pod "infra-operator-controller-manager-84b495f78-tgd62" (UID: "0be98506-68e8-438f-a46b-2225dd61b871") : secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:47 crc kubenswrapper[4966]: I1217 08:35:47.224839 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:47 crc kubenswrapper[4966]: E1217 08:35:47.225052 4966 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:47 crc kubenswrapper[4966]: E1217 08:35:47.225123 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert podName:add4eb97-35a5-4070-ac9f-e14eb5741aac nodeName:}" failed. No retries permitted until 2025-12-17 08:35:55.225104675 +0000 UTC m=+890.770174617 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert") pod "openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" (UID: "add4eb97-35a5-4070-ac9f-e14eb5741aac") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 17 08:35:47 crc kubenswrapper[4966]: I1217 08:35:47.326105 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:47 crc kubenswrapper[4966]: E1217 08:35:47.326268 4966 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 17 08:35:47 crc kubenswrapper[4966]: I1217 08:35:47.326326 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:47 crc kubenswrapper[4966]: E1217 08:35:47.326357 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:55.326331251 +0000 UTC m=+890.871401233 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "metrics-server-cert" not found Dec 17 08:35:47 crc kubenswrapper[4966]: E1217 08:35:47.326533 4966 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 17 08:35:47 crc kubenswrapper[4966]: E1217 08:35:47.326623 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:35:55.326604158 +0000 UTC m=+890.871674100 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "webhook-server-cert" not found Dec 17 08:35:53 crc kubenswrapper[4966]: E1217 08:35:53.947006 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 17 08:35:53 crc kubenswrapper[4966]: E1217 08:35:53.947594 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-79vbf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8665b56d78-mv4mc_openstack-operators(dd104116-5bb1-421b-ae58-9b7b6e6f79e5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:35:53 crc kubenswrapper[4966]: E1217 08:35:53.948777 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" podUID="dd104116-5bb1-421b-ae58-9b7b6e6f79e5" Dec 17 08:35:54 crc kubenswrapper[4966]: E1217 08:35:54.744587 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" podUID="dd104116-5bb1-421b-ae58-9b7b6e6f79e5" Dec 17 08:35:54 crc kubenswrapper[4966]: I1217 08:35:54.746988 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:35:54 crc kubenswrapper[4966]: E1217 08:35:54.747158 4966 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:54 crc kubenswrapper[4966]: E1217 08:35:54.747201 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert podName:0be98506-68e8-438f-a46b-2225dd61b871 nodeName:}" failed. No retries permitted until 2025-12-17 08:36:10.747186484 +0000 UTC m=+906.292256436 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert") pod "infra-operator-controller-manager-84b495f78-tgd62" (UID: "0be98506-68e8-438f-a46b-2225dd61b871") : secret "infra-operator-webhook-server-cert" not found Dec 17 08:35:55 crc kubenswrapper[4966]: I1217 08:35:55.255863 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:55 crc kubenswrapper[4966]: I1217 08:35:55.265645 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/add4eb97-35a5-4070-ac9f-e14eb5741aac-cert\") pod \"openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd\" (UID: \"add4eb97-35a5-4070-ac9f-e14eb5741aac\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:55 crc kubenswrapper[4966]: I1217 08:35:55.357575 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:55 crc kubenswrapper[4966]: I1217 08:35:55.357655 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:35:55 crc kubenswrapper[4966]: E1217 08:35:55.357800 4966 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 17 08:35:55 crc kubenswrapper[4966]: E1217 08:35:55.357858 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:36:11.357839937 +0000 UTC m=+906.902909889 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "metrics-server-cert" not found Dec 17 08:35:55 crc kubenswrapper[4966]: E1217 08:35:55.358309 4966 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 17 08:35:55 crc kubenswrapper[4966]: E1217 08:35:55.358352 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs podName:64c2ea0f-3e8c-4872-b130-df049355e147 nodeName:}" failed. No retries permitted until 2025-12-17 08:36:11.35834045 +0000 UTC m=+906.903410402 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs") pod "openstack-operator-controller-manager-5b4db5647c-2fdbb" (UID: "64c2ea0f-3e8c-4872-b130-df049355e147") : secret "webhook-server-cert" not found Dec 17 08:35:55 crc kubenswrapper[4966]: E1217 08:35:55.450472 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 17 08:35:55 crc kubenswrapper[4966]: E1217 08:35:55.450941 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rhd6g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6ccf486b9-64x68_openstack-operators(b288f5ca-8391-4157-8f02-886c88887df1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:35:55 crc kubenswrapper[4966]: E1217 08:35:55.452274 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" podUID="b288f5ca-8391-4157-8f02-886c88887df1" Dec 17 08:35:55 crc kubenswrapper[4966]: I1217 08:35:55.523816 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:35:55 crc kubenswrapper[4966]: E1217 08:35:55.751078 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" podUID="b288f5ca-8391-4157-8f02-886c88887df1" Dec 17 08:35:56 crc kubenswrapper[4966]: E1217 08:35:56.130599 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 17 08:35:56 crc kubenswrapper[4966]: E1217 08:35:56.130806 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gfmlg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7cd87b778f-b2q7b_openstack-operators(9c79b45a-796b-4bba-9daa-80d2c7e7d55b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:35:56 crc kubenswrapper[4966]: E1217 08:35:56.132275 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" podUID="9c79b45a-796b-4bba-9daa-80d2c7e7d55b" Dec 17 08:35:56 crc kubenswrapper[4966]: E1217 08:35:56.758204 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" podUID="9c79b45a-796b-4bba-9daa-80d2c7e7d55b" Dec 17 08:35:58 crc kubenswrapper[4966]: E1217 08:35:58.375538 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 17 08:35:58 crc kubenswrapper[4966]: E1217 08:35:58.376142 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cjmkq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-767f9d7567-g7zdx_openstack-operators(1916840c-817c-4ad4-849f-4eb0f4787a79): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:35:58 crc kubenswrapper[4966]: E1217 08:35:58.377284 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" podUID="1916840c-817c-4ad4-849f-4eb0f4787a79" Dec 17 08:35:58 crc kubenswrapper[4966]: E1217 08:35:58.768685 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027\\\"\"" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" podUID="1916840c-817c-4ad4-849f-4eb0f4787a79" Dec 17 08:35:58 crc kubenswrapper[4966]: I1217 08:35:58.878616 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4fzss"] Dec 17 08:35:58 crc kubenswrapper[4966]: I1217 08:35:58.880148 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:58 crc kubenswrapper[4966]: I1217 08:35:58.887390 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4fzss"] Dec 17 08:35:58 crc kubenswrapper[4966]: I1217 08:35:58.920749 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-utilities\") pod \"redhat-marketplace-4fzss\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:58 crc kubenswrapper[4966]: I1217 08:35:58.920809 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-catalog-content\") pod \"redhat-marketplace-4fzss\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:58 crc kubenswrapper[4966]: I1217 08:35:58.920889 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smdqc\" (UniqueName: \"kubernetes.io/projected/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-kube-api-access-smdqc\") pod \"redhat-marketplace-4fzss\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:59 crc kubenswrapper[4966]: I1217 08:35:59.022526 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-utilities\") pod \"redhat-marketplace-4fzss\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:59 crc kubenswrapper[4966]: I1217 08:35:59.022811 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-catalog-content\") pod \"redhat-marketplace-4fzss\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:59 crc kubenswrapper[4966]: I1217 08:35:59.022938 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smdqc\" (UniqueName: \"kubernetes.io/projected/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-kube-api-access-smdqc\") pod \"redhat-marketplace-4fzss\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:59 crc kubenswrapper[4966]: I1217 08:35:59.023810 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-utilities\") pod \"redhat-marketplace-4fzss\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:59 crc kubenswrapper[4966]: I1217 08:35:59.024277 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-catalog-content\") pod \"redhat-marketplace-4fzss\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:59 crc kubenswrapper[4966]: I1217 08:35:59.043254 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smdqc\" (UniqueName: \"kubernetes.io/projected/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-kube-api-access-smdqc\") pod \"redhat-marketplace-4fzss\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:59 crc kubenswrapper[4966]: E1217 08:35:59.118168 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 17 08:35:59 crc kubenswrapper[4966]: E1217 08:35:59.118368 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7rbgc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-59b8dcb766-w6wjs_openstack-operators(2803b782-fe84-4912-b7e6-528240fe14cb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:35:59 crc kubenswrapper[4966]: E1217 08:35:59.119700 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" podUID="2803b782-fe84-4912-b7e6-528240fe14cb" Dec 17 08:35:59 crc kubenswrapper[4966]: I1217 08:35:59.203065 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:35:59 crc kubenswrapper[4966]: E1217 08:35:59.778157 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\"" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" podUID="2803b782-fe84-4912-b7e6-528240fe14cb" Dec 17 08:36:00 crc kubenswrapper[4966]: E1217 08:36:00.605007 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 17 08:36:00 crc kubenswrapper[4966]: E1217 08:36:00.605262 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7g2qv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5fbbf8b6cc-6z6kr_openstack-operators(ae9e23d6-8355-49f3-8558-1dcc50b0d12e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:00 crc kubenswrapper[4966]: E1217 08:36:00.607913 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" podUID="ae9e23d6-8355-49f3-8558-1dcc50b0d12e" Dec 17 08:36:00 crc kubenswrapper[4966]: E1217 08:36:00.784271 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" podUID="ae9e23d6-8355-49f3-8558-1dcc50b0d12e" Dec 17 08:36:01 crc kubenswrapper[4966]: E1217 08:36:01.093601 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Dec 17 08:36:01 crc kubenswrapper[4966]: E1217 08:36:01.094028 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c4gct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-66f8b87655-l6ch2_openstack-operators(85b3221d-40ef-454b-920f-f0934adb7af1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:01 crc kubenswrapper[4966]: E1217 08:36:01.095463 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" podUID="85b3221d-40ef-454b-920f-f0934adb7af1" Dec 17 08:36:01 crc kubenswrapper[4966]: E1217 08:36:01.792745 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a\\\"\"" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" podUID="85b3221d-40ef-454b-920f-f0934adb7af1" Dec 17 08:36:03 crc kubenswrapper[4966]: E1217 08:36:03.367654 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 17 08:36:03 crc kubenswrapper[4966]: E1217 08:36:03.368001 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hggzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-f76f4954c-mfghn_openstack-operators(8157370d-5a0a-4d55-98be-1061654d3a37): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:03 crc kubenswrapper[4966]: E1217 08:36:03.369264 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" podUID="8157370d-5a0a-4d55-98be-1061654d3a37" Dec 17 08:36:03 crc kubenswrapper[4966]: E1217 08:36:03.804723 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" podUID="8157370d-5a0a-4d55-98be-1061654d3a37" Dec 17 08:36:04 crc kubenswrapper[4966]: E1217 08:36:04.097521 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f" Dec 17 08:36:04 crc kubenswrapper[4966]: E1217 08:36:04.097743 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wqj2s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-97d456b9-g6xkf_openstack-operators(1458fa23-50b0-4dad-bed4-d9859806d337): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:04 crc kubenswrapper[4966]: E1217 08:36:04.099289 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" podUID="1458fa23-50b0-4dad-bed4-d9859806d337" Dec 17 08:36:04 crc kubenswrapper[4966]: E1217 08:36:04.808361 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87" Dec 17 08:36:04 crc kubenswrapper[4966]: E1217 08:36:04.808769 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rz2gn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-f458558d7-f98m5_openstack-operators(9e86820d-0103-4900-b92a-ef86d1292689): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:04 crc kubenswrapper[4966]: E1217 08:36:04.810036 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" podUID="9e86820d-0103-4900-b92a-ef86d1292689" Dec 17 08:36:04 crc kubenswrapper[4966]: E1217 08:36:04.818032 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" podUID="1458fa23-50b0-4dad-bed4-d9859806d337" Dec 17 08:36:05 crc kubenswrapper[4966]: E1217 08:36:05.824904 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" podUID="9e86820d-0103-4900-b92a-ef86d1292689" Dec 17 08:36:07 crc kubenswrapper[4966]: I1217 08:36:07.993291 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-glcd5"] Dec 17 08:36:07 crc kubenswrapper[4966]: I1217 08:36:07.995651 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.006996 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-glcd5"] Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.166447 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-catalog-content\") pod \"redhat-operators-glcd5\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.167087 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkrtf\" (UniqueName: \"kubernetes.io/projected/8d2c4ece-23cd-4941-9a72-9958ee533c09-kube-api-access-bkrtf\") pod \"redhat-operators-glcd5\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.167160 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-utilities\") pod \"redhat-operators-glcd5\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: E1217 08:36:08.255985 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 17 08:36:08 crc kubenswrapper[4966]: E1217 08:36:08.256231 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-brngw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-68c649d9d-cr8zh_openstack-operators(62fbcfd8-be0b-4b98-bc08-337173ee2758): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:08 crc kubenswrapper[4966]: E1217 08:36:08.257572 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" podUID="62fbcfd8-be0b-4b98-bc08-337173ee2758" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.269113 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkrtf\" (UniqueName: \"kubernetes.io/projected/8d2c4ece-23cd-4941-9a72-9958ee533c09-kube-api-access-bkrtf\") pod \"redhat-operators-glcd5\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.269367 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-utilities\") pod \"redhat-operators-glcd5\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.269502 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-catalog-content\") pod \"redhat-operators-glcd5\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.270257 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-catalog-content\") pod \"redhat-operators-glcd5\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.271988 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-utilities\") pod \"redhat-operators-glcd5\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.290617 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkrtf\" (UniqueName: \"kubernetes.io/projected/8d2c4ece-23cd-4941-9a72-9958ee533c09-kube-api-access-bkrtf\") pod \"redhat-operators-glcd5\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: I1217 08:36:08.312922 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:08 crc kubenswrapper[4966]: E1217 08:36:08.849491 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" podUID="62fbcfd8-be0b-4b98-bc08-337173ee2758" Dec 17 08:36:09 crc kubenswrapper[4966]: E1217 08:36:09.368621 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 17 08:36:09 crc kubenswrapper[4966]: E1217 08:36:09.368775 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lbtjh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5c6df8f9-rcsdh_openstack-operators(2cc894eb-b898-4c2e-b379-7a48d08d6293): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:09 crc kubenswrapper[4966]: E1217 08:36:09.370006 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" podUID="2cc894eb-b898-4c2e-b379-7a48d08d6293" Dec 17 08:36:09 crc kubenswrapper[4966]: E1217 08:36:09.858005 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" podUID="2cc894eb-b898-4c2e-b379-7a48d08d6293" Dec 17 08:36:10 crc kubenswrapper[4966]: E1217 08:36:10.039207 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 17 08:36:10 crc kubenswrapper[4966]: E1217 08:36:10.039439 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmtvz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-5c7cbf548f-zmxmt_openstack-operators(c1229400-ebf1-4fb6-a8a0-4f7292c22dfc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:10 crc kubenswrapper[4966]: E1217 08:36:10.043900 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" podUID="c1229400-ebf1-4fb6-a8a0-4f7292c22dfc" Dec 17 08:36:10 crc kubenswrapper[4966]: E1217 08:36:10.598116 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a" Dec 17 08:36:10 crc kubenswrapper[4966]: E1217 08:36:10.598261 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6x9p7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-55f78b7c4c-z9pzc_openstack-operators(861f02ab-0025-4dbb-b020-e02561b3830e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:10 crc kubenswrapper[4966]: E1217 08:36:10.599447 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" podUID="861f02ab-0025-4dbb-b020-e02561b3830e" Dec 17 08:36:10 crc kubenswrapper[4966]: I1217 08:36:10.805040 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:36:10 crc kubenswrapper[4966]: I1217 08:36:10.812622 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0be98506-68e8-438f-a46b-2225dd61b871-cert\") pod \"infra-operator-controller-manager-84b495f78-tgd62\" (UID: \"0be98506-68e8-438f-a46b-2225dd61b871\") " pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:36:10 crc kubenswrapper[4966]: E1217 08:36:10.863521 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" podUID="c1229400-ebf1-4fb6-a8a0-4f7292c22dfc" Dec 17 08:36:10 crc kubenswrapper[4966]: I1217 08:36:10.997904 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-8l8k8" Dec 17 08:36:11 crc kubenswrapper[4966]: I1217 08:36:11.006628 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:36:11 crc kubenswrapper[4966]: E1217 08:36:11.188684 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 17 08:36:11 crc kubenswrapper[4966]: E1217 08:36:11.188909 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kvfvf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-756ccf86c7-ksgt7_openstack-operators(5277f562-0c42-4ca8-b8a5-aa345d02fca7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:36:11 crc kubenswrapper[4966]: E1217 08:36:11.190246 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" podUID="5277f562-0c42-4ca8-b8a5-aa345d02fca7" Dec 17 08:36:11 crc kubenswrapper[4966]: I1217 08:36:11.413834 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:36:11 crc kubenswrapper[4966]: I1217 08:36:11.413935 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:36:11 crc kubenswrapper[4966]: I1217 08:36:11.430794 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-metrics-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:36:11 crc kubenswrapper[4966]: I1217 08:36:11.432313 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/64c2ea0f-3e8c-4872-b130-df049355e147-webhook-certs\") pod \"openstack-operator-controller-manager-5b4db5647c-2fdbb\" (UID: \"64c2ea0f-3e8c-4872-b130-df049355e147\") " pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:36:11 crc kubenswrapper[4966]: I1217 08:36:11.713956 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-m64k2" Dec 17 08:36:11 crc kubenswrapper[4966]: I1217 08:36:11.722700 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.457935 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd"] Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.631785 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4fzss"] Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.777705 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-glcd5"] Dec 17 08:36:13 crc kubenswrapper[4966]: W1217 08:36:13.804560 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d2c4ece_23cd_4941_9a72_9958ee533c09.slice/crio-e7082b97031349cd4983b0ba83f5916681f18fab25677754d6206b847a1a5018 WatchSource:0}: Error finding container e7082b97031349cd4983b0ba83f5916681f18fab25677754d6206b847a1a5018: Status 404 returned error can't find the container with id e7082b97031349cd4983b0ba83f5916681f18fab25677754d6206b847a1a5018 Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.900572 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" event={"ID":"202f0a3b-a9da-4ce9-9297-10cba922d8ab","Type":"ContainerStarted","Data":"9c3d2f16157e974d355989de7b5f4eb79a7f6ab5d837f853236a565bd04db78a"} Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.902359 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.910886 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" event={"ID":"1916840c-817c-4ad4-849f-4eb0f4787a79","Type":"ContainerStarted","Data":"24cf3bf8e67b115d310808f44e1d4a56a78f939cf5c55233251fa11a496d61ab"} Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.911565 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.914594 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84b495f78-tgd62"] Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.920131 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" event={"ID":"dd104116-5bb1-421b-ae58-9b7b6e6f79e5","Type":"ContainerStarted","Data":"240a0cbbc14539e80d60f3cce41c2598f117a95ec2c0f2f6c9888985b3cee73e"} Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.920981 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.931541 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" event={"ID":"add4eb97-35a5-4070-ac9f-e14eb5741aac","Type":"ContainerStarted","Data":"4c5fc71bc2bdd75cd3d659b2846c86839d55a47b5ebd15bb38de1fdcf3eac706"} Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.953753 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" podStartSLOduration=5.242571431 podStartE2EDuration="35.953730522s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:39.873273742 +0000 UTC m=+875.418343684" lastFinishedPulling="2025-12-17 08:36:10.584432833 +0000 UTC m=+906.129502775" observedRunningTime="2025-12-17 08:36:13.94202439 +0000 UTC m=+909.487094352" watchObservedRunningTime="2025-12-17 08:36:13.953730522 +0000 UTC m=+909.498800464" Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.957077 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" event={"ID":"cfe05261-97ae-4fd8-b178-1a0679c689da","Type":"ContainerStarted","Data":"32217e3234d27d8d34eb12e9552e2bb6153890ce6317169fe4b8c8273f9b98e8"} Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.958000 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.965209 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-glcd5" event={"ID":"8d2c4ece-23cd-4941-9a72-9958ee533c09","Type":"ContainerStarted","Data":"e7082b97031349cd4983b0ba83f5916681f18fab25677754d6206b847a1a5018"} Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.971802 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" event={"ID":"d412de43-e6c6-4c8a-a947-3811cad9d6a8","Type":"ContainerStarted","Data":"85c9d6150312871631b4b352f0c04973651044902787cdb7fe84e9660456e628"} Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.972406 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.976979 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795j5" event={"ID":"057025a8-6d42-4a30-8156-b041e8c4d9aa","Type":"ContainerStarted","Data":"efa8eb39af70f49a13d4f26ebb81dfe84ded7dca970249bd8725ad47b0b70921"} Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.978958 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" podStartSLOduration=4.291627495 podStartE2EDuration="35.978940866s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.431171459 +0000 UTC m=+876.976241401" lastFinishedPulling="2025-12-17 08:36:13.11848483 +0000 UTC m=+908.663554772" observedRunningTime="2025-12-17 08:36:13.974031511 +0000 UTC m=+909.519101453" watchObservedRunningTime="2025-12-17 08:36:13.978940866 +0000 UTC m=+909.524010808" Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.981919 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4fzss" event={"ID":"5f81a852-06ec-405a-ba4e-e68b12b6e5e0","Type":"ContainerStarted","Data":"a58ee74eac10496c5dccac4b6479a0d2371edb5d1f34014873d62fb784c20def"} Dec 17 08:36:13 crc kubenswrapper[4966]: I1217 08:36:13.993094 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb"] Dec 17 08:36:14 crc kubenswrapper[4966]: I1217 08:36:14.025043 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" podStartSLOduration=3.10486858 podStartE2EDuration="36.025025734s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:40.208328011 +0000 UTC m=+875.753397953" lastFinishedPulling="2025-12-17 08:36:13.128485155 +0000 UTC m=+908.673555107" observedRunningTime="2025-12-17 08:36:14.012274073 +0000 UTC m=+909.557344035" watchObservedRunningTime="2025-12-17 08:36:14.025025734 +0000 UTC m=+909.570095676" Dec 17 08:36:14 crc kubenswrapper[4966]: I1217 08:36:14.089389 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" podStartSLOduration=6.886272539 podStartE2EDuration="36.089368634s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.382335135 +0000 UTC m=+876.927405077" lastFinishedPulling="2025-12-17 08:36:10.58543124 +0000 UTC m=+906.130501172" observedRunningTime="2025-12-17 08:36:14.068089689 +0000 UTC m=+909.613159641" watchObservedRunningTime="2025-12-17 08:36:14.089368634 +0000 UTC m=+909.634438576" Dec 17 08:36:14 crc kubenswrapper[4966]: I1217 08:36:14.121528 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" podStartSLOduration=5.8610542500000005 podStartE2EDuration="36.121511039s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:40.933924337 +0000 UTC m=+876.478994279" lastFinishedPulling="2025-12-17 08:36:11.194381126 +0000 UTC m=+906.739451068" observedRunningTime="2025-12-17 08:36:14.115546535 +0000 UTC m=+909.660616487" watchObservedRunningTime="2025-12-17 08:36:14.121511039 +0000 UTC m=+909.666580981" Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.016536 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" event={"ID":"64c2ea0f-3e8c-4872-b130-df049355e147","Type":"ContainerStarted","Data":"adaf7d1f3e0e439248ce7f1830f83884e67563af739c3d17cd9a4cacead5b110"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.016820 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" event={"ID":"64c2ea0f-3e8c-4872-b130-df049355e147","Type":"ContainerStarted","Data":"ca679deedb2042f390c465dc9a8a711c9164731fc7a967c3ec48123bd469b760"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.018276 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.030181 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" event={"ID":"9c79b45a-796b-4bba-9daa-80d2c7e7d55b","Type":"ContainerStarted","Data":"bb843b2e7e84fb961fb4fafd0c96967eb83dfc97d6c99c3f7f8849c2813df8bb"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.030738 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.039196 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" event={"ID":"2eb3f60e-624a-43e1-a27a-48d466689c78","Type":"ContainerStarted","Data":"591aff21bd5431b9e6d7772e9114a92a1e859b0206899ff7db3add0469093a22"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.039508 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.041735 4966 generic.go:334] "Generic (PLEG): container finished" podID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerID="d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194" exitCode=0 Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.041866 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-glcd5" event={"ID":"8d2c4ece-23cd-4941-9a72-9958ee533c09","Type":"ContainerDied","Data":"d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.048148 4966 generic.go:334] "Generic (PLEG): container finished" podID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerID="efa8eb39af70f49a13d4f26ebb81dfe84ded7dca970249bd8725ad47b0b70921" exitCode=0 Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.048217 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795j5" event={"ID":"057025a8-6d42-4a30-8156-b041e8c4d9aa","Type":"ContainerDied","Data":"efa8eb39af70f49a13d4f26ebb81dfe84ded7dca970249bd8725ad47b0b70921"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.053315 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" event={"ID":"0be98506-68e8-438f-a46b-2225dd61b871","Type":"ContainerStarted","Data":"e1b3ebfe33376064c1c2b03fc8cc8967e86382fc060a05c043d7bd1256a42998"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.058445 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" podStartSLOduration=36.058434409 podStartE2EDuration="36.058434409s" podCreationTimestamp="2025-12-17 08:35:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:36:15.057090782 +0000 UTC m=+910.602160724" watchObservedRunningTime="2025-12-17 08:36:15.058434409 +0000 UTC m=+910.603504351" Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.060771 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" event={"ID":"f485453d-7629-41e7-a1c7-2c0b284bb551","Type":"ContainerStarted","Data":"d2f836d8760187a494f12716b6314d5496c5256d9e3cc0f6b2976fb7bed29368"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.071792 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" event={"ID":"b288f5ca-8391-4157-8f02-886c88887df1","Type":"ContainerStarted","Data":"cc2a3fd01f69fdf9025396fe05f3ce2cb7cbc88ccafe7c2e7a0d57ee7a54e4a1"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.072423 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.085304 4966 generic.go:334] "Generic (PLEG): container finished" podID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerID="32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972" exitCode=0 Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.085654 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4fzss" event={"ID":"5f81a852-06ec-405a-ba4e-e68b12b6e5e0","Type":"ContainerDied","Data":"32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972"} Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.119718 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" podStartSLOduration=5.042452645 podStartE2EDuration="37.119701135s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:40.208026113 +0000 UTC m=+875.753096055" lastFinishedPulling="2025-12-17 08:36:12.285274603 +0000 UTC m=+907.830344545" observedRunningTime="2025-12-17 08:36:15.11005543 +0000 UTC m=+910.655125382" watchObservedRunningTime="2025-12-17 08:36:15.119701135 +0000 UTC m=+910.664771077" Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.138138 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" podStartSLOduration=5.038611249 podStartE2EDuration="37.138121692s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.026556905 +0000 UTC m=+876.571626847" lastFinishedPulling="2025-12-17 08:36:13.126067338 +0000 UTC m=+908.671137290" observedRunningTime="2025-12-17 08:36:15.135722616 +0000 UTC m=+910.680792558" watchObservedRunningTime="2025-12-17 08:36:15.138121692 +0000 UTC m=+910.683191634" Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.219073 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" podStartSLOduration=4.380827699 podStartE2EDuration="37.219054279s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:40.358575045 +0000 UTC m=+875.903644987" lastFinishedPulling="2025-12-17 08:36:13.196801625 +0000 UTC m=+908.741871567" observedRunningTime="2025-12-17 08:36:15.218284188 +0000 UTC m=+910.763354130" watchObservedRunningTime="2025-12-17 08:36:15.219054279 +0000 UTC m=+910.764124221" Dec 17 08:36:15 crc kubenswrapper[4966]: I1217 08:36:15.219200 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8fvcq" podStartSLOduration=4.589065236 podStartE2EDuration="36.219195713s" podCreationTimestamp="2025-12-17 08:35:39 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.587364977 +0000 UTC m=+877.132434919" lastFinishedPulling="2025-12-17 08:36:13.217495464 +0000 UTC m=+908.762565396" observedRunningTime="2025-12-17 08:36:15.191600253 +0000 UTC m=+910.736670205" watchObservedRunningTime="2025-12-17 08:36:15.219195713 +0000 UTC m=+910.764265655" Dec 17 08:36:18 crc kubenswrapper[4966]: I1217 08:36:18.799040 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-95949466-p69c8" Dec 17 08:36:18 crc kubenswrapper[4966]: I1217 08:36:18.879264 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-767f9d7567-g7zdx" Dec 17 08:36:19 crc kubenswrapper[4966]: I1217 08:36:19.272327 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5fdd9786f7-tv6n9" Dec 17 08:36:19 crc kubenswrapper[4966]: I1217 08:36:19.511182 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-b2q7b" Dec 17 08:36:19 crc kubenswrapper[4966]: I1217 08:36:19.728098 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-fj2v2" Dec 17 08:36:19 crc kubenswrapper[4966]: I1217 08:36:19.796179 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8665b56d78-mv4mc" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.123050 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" event={"ID":"2803b782-fe84-4912-b7e6-528240fe14cb","Type":"ContainerStarted","Data":"82038d3c64bb311a6cec26d1434123797bb65175f6f2014a88d633b9310be7a8"} Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.123264 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.126489 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795j5" event={"ID":"057025a8-6d42-4a30-8156-b041e8c4d9aa","Type":"ContainerStarted","Data":"9360b2c3dc58d21759bb96025da73cb17cd208ea4d9e5918c47555a440fccd99"} Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.128702 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" event={"ID":"8157370d-5a0a-4d55-98be-1061654d3a37","Type":"ContainerStarted","Data":"0bf82b5ee26284872d6b9f1c1170c30f34d0289ee5e222359f34b95a2c44e119"} Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.128919 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.132362 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4fzss" event={"ID":"5f81a852-06ec-405a-ba4e-e68b12b6e5e0","Type":"ContainerStarted","Data":"9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5"} Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.134089 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" event={"ID":"85b3221d-40ef-454b-920f-f0934adb7af1","Type":"ContainerStarted","Data":"cbe404fd32f30a4aa8d869d48f853fd0118eaf586b410fad7e8692576b59f231"} Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.134457 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.136736 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" event={"ID":"ae9e23d6-8355-49f3-8558-1dcc50b0d12e","Type":"ContainerStarted","Data":"4aa43abbcf875113638abd42831c5116f5cf92f41290a0b22e4c3546f0bbc279"} Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.136911 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.141524 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-glcd5" event={"ID":"8d2c4ece-23cd-4941-9a72-9958ee533c09","Type":"ContainerStarted","Data":"4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c"} Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.147371 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" podStartSLOduration=3.681531678 podStartE2EDuration="42.147354907s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:40.569385876 +0000 UTC m=+876.114455818" lastFinishedPulling="2025-12-17 08:36:19.035209095 +0000 UTC m=+914.580279047" observedRunningTime="2025-12-17 08:36:20.140710344 +0000 UTC m=+915.685780286" watchObservedRunningTime="2025-12-17 08:36:20.147354907 +0000 UTC m=+915.692424849" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.170944 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" podStartSLOduration=3.470695567 podStartE2EDuration="42.170929146s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:40.559929456 +0000 UTC m=+876.104999398" lastFinishedPulling="2025-12-17 08:36:19.260163035 +0000 UTC m=+914.805232977" observedRunningTime="2025-12-17 08:36:20.165397654 +0000 UTC m=+915.710467596" watchObservedRunningTime="2025-12-17 08:36:20.170929146 +0000 UTC m=+915.715999088" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.188440 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-795j5" podStartSLOduration=3.4963037630000002 podStartE2EDuration="40.188395596s" podCreationTimestamp="2025-12-17 08:35:40 +0000 UTC" firstStartedPulling="2025-12-17 08:35:42.611278351 +0000 UTC m=+878.156348293" lastFinishedPulling="2025-12-17 08:36:19.303370174 +0000 UTC m=+914.848440126" observedRunningTime="2025-12-17 08:36:20.180622623 +0000 UTC m=+915.725692575" watchObservedRunningTime="2025-12-17 08:36:20.188395596 +0000 UTC m=+915.733465538" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.227853 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" podStartSLOduration=3.531950441 podStartE2EDuration="42.227837381s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:40.961639969 +0000 UTC m=+876.506709911" lastFinishedPulling="2025-12-17 08:36:19.657526909 +0000 UTC m=+915.202596851" observedRunningTime="2025-12-17 08:36:20.222242807 +0000 UTC m=+915.767312749" watchObservedRunningTime="2025-12-17 08:36:20.227837381 +0000 UTC m=+915.772907323" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.283270 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" podStartSLOduration=4.646628944 podStartE2EDuration="42.283252726s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.396488775 +0000 UTC m=+876.941558717" lastFinishedPulling="2025-12-17 08:36:19.033112557 +0000 UTC m=+914.578182499" observedRunningTime="2025-12-17 08:36:20.275561154 +0000 UTC m=+915.820631096" watchObservedRunningTime="2025-12-17 08:36:20.283252726 +0000 UTC m=+915.828322668" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.679889 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:36:20 crc kubenswrapper[4966]: I1217 08:36:20.679960 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:36:21 crc kubenswrapper[4966]: I1217 08:36:21.151021 4966 generic.go:334] "Generic (PLEG): container finished" podID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerID="4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c" exitCode=0 Dec 17 08:36:21 crc kubenswrapper[4966]: I1217 08:36:21.151098 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-glcd5" event={"ID":"8d2c4ece-23cd-4941-9a72-9958ee533c09","Type":"ContainerDied","Data":"4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c"} Dec 17 08:36:21 crc kubenswrapper[4966]: I1217 08:36:21.154470 4966 generic.go:334] "Generic (PLEG): container finished" podID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerID="9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5" exitCode=0 Dec 17 08:36:21 crc kubenswrapper[4966]: I1217 08:36:21.155087 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4fzss" event={"ID":"5f81a852-06ec-405a-ba4e-e68b12b6e5e0","Type":"ContainerDied","Data":"9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5"} Dec 17 08:36:21 crc kubenswrapper[4966]: I1217 08:36:21.729693 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5b4db5647c-2fdbb" Dec 17 08:36:21 crc kubenswrapper[4966]: I1217 08:36:21.750040 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-795j5" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="registry-server" probeResult="failure" output=< Dec 17 08:36:21 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:36:21 crc kubenswrapper[4966]: > Dec 17 08:36:22 crc kubenswrapper[4966]: E1217 08:36:22.834288 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\"" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" podUID="5277f562-0c42-4ca8-b8a5-aa345d02fca7" Dec 17 08:36:23 crc kubenswrapper[4966]: E1217 08:36:23.833002 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" podUID="861f02ab-0025-4dbb-b020-e02561b3830e" Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.857486 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bvd4j"] Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.859209 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.883621 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k2wk\" (UniqueName: \"kubernetes.io/projected/55f1659f-c016-4621-825a-be04300e0bc3-kube-api-access-8k2wk\") pod \"community-operators-bvd4j\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.883732 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-catalog-content\") pod \"community-operators-bvd4j\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.883787 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-utilities\") pod \"community-operators-bvd4j\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.945712 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bvd4j"] Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.985904 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-utilities\") pod \"community-operators-bvd4j\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.986202 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k2wk\" (UniqueName: \"kubernetes.io/projected/55f1659f-c016-4621-825a-be04300e0bc3-kube-api-access-8k2wk\") pod \"community-operators-bvd4j\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.986248 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-catalog-content\") pod \"community-operators-bvd4j\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.986646 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-catalog-content\") pod \"community-operators-bvd4j\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:25 crc kubenswrapper[4966]: I1217 08:36:25.986852 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-utilities\") pod \"community-operators-bvd4j\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.022141 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k2wk\" (UniqueName: \"kubernetes.io/projected/55f1659f-c016-4621-825a-be04300e0bc3-kube-api-access-8k2wk\") pod \"community-operators-bvd4j\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.174707 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.229493 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" event={"ID":"c1229400-ebf1-4fb6-a8a0-4f7292c22dfc","Type":"ContainerStarted","Data":"b708cee3b7d38e5ebaab38de2dabf63c09f00d0252b5cad52b2bc6302565ff51"} Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.230628 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.232286 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" event={"ID":"1458fa23-50b0-4dad-bed4-d9859806d337","Type":"ContainerStarted","Data":"713630c8d6fee8e4adc7416c34e5cb859adf19eb11e4ececc396804e7fcebd8a"} Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.232623 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.243277 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" event={"ID":"2cc894eb-b898-4c2e-b379-7a48d08d6293","Type":"ContainerStarted","Data":"f813bb156b389b68b8c31fe5bc910318eaaabdb8581b6224a83a2366d8d853a6"} Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.243512 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.248327 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4fzss" event={"ID":"5f81a852-06ec-405a-ba4e-e68b12b6e5e0","Type":"ContainerStarted","Data":"f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa"} Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.253989 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-glcd5" event={"ID":"8d2c4ece-23cd-4941-9a72-9958ee533c09","Type":"ContainerStarted","Data":"64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099"} Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.255851 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" event={"ID":"9e86820d-0103-4900-b92a-ef86d1292689","Type":"ContainerStarted","Data":"0b07772229f132d6a9edb76fa5010ed0aa4f9e07475e694b576e46639ebb1168"} Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.256212 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.262283 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" podStartSLOduration=4.180463655 podStartE2EDuration="48.262268085s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:40.941909956 +0000 UTC m=+876.486979898" lastFinishedPulling="2025-12-17 08:36:25.023714386 +0000 UTC m=+920.568784328" observedRunningTime="2025-12-17 08:36:26.25663312 +0000 UTC m=+921.801703062" watchObservedRunningTime="2025-12-17 08:36:26.262268085 +0000 UTC m=+921.807338027" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.264138 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" event={"ID":"0be98506-68e8-438f-a46b-2225dd61b871","Type":"ContainerStarted","Data":"d12344eb7ce7ffc0d2722e141389be59ad6c68a63f692b37bc6ea3f785f0cf88"} Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.264241 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.287475 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" event={"ID":"add4eb97-35a5-4070-ac9f-e14eb5741aac","Type":"ContainerStarted","Data":"5a027b55f13b2b7fc6a65ca69ddaee25244d71a8d6b175c1906d005aea767ea3"} Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.287834 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.302073 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" event={"ID":"62fbcfd8-be0b-4b98-bc08-337173ee2758","Type":"ContainerStarted","Data":"9670bf072c3ab3aef3895af471442cfbb3ff7e0d126222f9e5d793855b5f4184"} Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.302733 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.340134 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4fzss" podStartSLOduration=18.36933229 podStartE2EDuration="28.340118828s" podCreationTimestamp="2025-12-17 08:35:58 +0000 UTC" firstStartedPulling="2025-12-17 08:36:15.088012723 +0000 UTC m=+910.633082665" lastFinishedPulling="2025-12-17 08:36:25.058799261 +0000 UTC m=+920.603869203" observedRunningTime="2025-12-17 08:36:26.323138371 +0000 UTC m=+921.868208313" watchObservedRunningTime="2025-12-17 08:36:26.340118828 +0000 UTC m=+921.885188770" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.433726 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" podStartSLOduration=4.829286001 podStartE2EDuration="48.433710043s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.400721112 +0000 UTC m=+876.945791064" lastFinishedPulling="2025-12-17 08:36:25.005145154 +0000 UTC m=+920.550215106" observedRunningTime="2025-12-17 08:36:26.386693039 +0000 UTC m=+921.931762981" watchObservedRunningTime="2025-12-17 08:36:26.433710043 +0000 UTC m=+921.978779985" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.471037 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" podStartSLOduration=4.155279833 podStartE2EDuration="48.471022259s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.205586612 +0000 UTC m=+876.750656554" lastFinishedPulling="2025-12-17 08:36:25.521329038 +0000 UTC m=+921.066398980" observedRunningTime="2025-12-17 08:36:26.431744968 +0000 UTC m=+921.976814910" watchObservedRunningTime="2025-12-17 08:36:26.471022259 +0000 UTC m=+922.016092201" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.523225 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" podStartSLOduration=4.326381122 podStartE2EDuration="48.523206656s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:40.856913407 +0000 UTC m=+876.401983349" lastFinishedPulling="2025-12-17 08:36:25.053738941 +0000 UTC m=+920.598808883" observedRunningTime="2025-12-17 08:36:26.472204852 +0000 UTC m=+922.017274784" watchObservedRunningTime="2025-12-17 08:36:26.523206656 +0000 UTC m=+922.068276598" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.526011 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-glcd5" podStartSLOduration=9.564163683 podStartE2EDuration="19.526001003s" podCreationTimestamp="2025-12-17 08:36:07 +0000 UTC" firstStartedPulling="2025-12-17 08:36:15.043556101 +0000 UTC m=+910.588626043" lastFinishedPulling="2025-12-17 08:36:25.005393421 +0000 UTC m=+920.550463363" observedRunningTime="2025-12-17 08:36:26.522087595 +0000 UTC m=+922.067157547" watchObservedRunningTime="2025-12-17 08:36:26.526001003 +0000 UTC m=+922.071070935" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.613999 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" podStartSLOduration=37.475220857 podStartE2EDuration="48.613981023s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:36:13.931013377 +0000 UTC m=+909.476083319" lastFinishedPulling="2025-12-17 08:36:25.069773543 +0000 UTC m=+920.614843485" observedRunningTime="2025-12-17 08:36:26.579676849 +0000 UTC m=+922.124746791" watchObservedRunningTime="2025-12-17 08:36:26.613981023 +0000 UTC m=+922.159050965" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.680644 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" podStartSLOduration=37.241778895 podStartE2EDuration="48.680628107s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:36:13.63091595 +0000 UTC m=+909.175985892" lastFinishedPulling="2025-12-17 08:36:25.069765162 +0000 UTC m=+920.614835104" observedRunningTime="2025-12-17 08:36:26.672217516 +0000 UTC m=+922.217287458" watchObservedRunningTime="2025-12-17 08:36:26.680628107 +0000 UTC m=+922.225698049" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.722222 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" podStartSLOduration=4.8987207900000005 podStartE2EDuration="48.722206131s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.246691163 +0000 UTC m=+876.791761105" lastFinishedPulling="2025-12-17 08:36:25.070176504 +0000 UTC m=+920.615246446" observedRunningTime="2025-12-17 08:36:26.721540922 +0000 UTC m=+922.266610864" watchObservedRunningTime="2025-12-17 08:36:26.722206131 +0000 UTC m=+922.267276073" Dec 17 08:36:26 crc kubenswrapper[4966]: I1217 08:36:26.993384 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bvd4j"] Dec 17 08:36:27 crc kubenswrapper[4966]: I1217 08:36:27.309862 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvd4j" event={"ID":"55f1659f-c016-4621-825a-be04300e0bc3","Type":"ContainerStarted","Data":"c9982a4d35cbba486de2e655e415d7b881869c865bb25417475fe8efebd7ebb5"} Dec 17 08:36:28 crc kubenswrapper[4966]: I1217 08:36:28.313682 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:28 crc kubenswrapper[4966]: I1217 08:36:28.313730 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:28 crc kubenswrapper[4966]: I1217 08:36:28.321167 4966 generic.go:334] "Generic (PLEG): container finished" podID="55f1659f-c016-4621-825a-be04300e0bc3" containerID="40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756" exitCode=0 Dec 17 08:36:28 crc kubenswrapper[4966]: I1217 08:36:28.322375 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvd4j" event={"ID":"55f1659f-c016-4621-825a-be04300e0bc3","Type":"ContainerDied","Data":"40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756"} Dec 17 08:36:28 crc kubenswrapper[4966]: I1217 08:36:28.773734 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5f98b4754f-dp5bm" Dec 17 08:36:28 crc kubenswrapper[4966]: I1217 08:36:28.885405 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-l6ch2" Dec 17 08:36:28 crc kubenswrapper[4966]: I1217 08:36:28.957474 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-59b8dcb766-w6wjs" Dec 17 08:36:28 crc kubenswrapper[4966]: I1217 08:36:28.991180 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6ccf486b9-64x68" Dec 17 08:36:29 crc kubenswrapper[4966]: I1217 08:36:29.204308 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:36:29 crc kubenswrapper[4966]: I1217 08:36:29.204366 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:36:29 crc kubenswrapper[4966]: I1217 08:36:29.332733 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvd4j" event={"ID":"55f1659f-c016-4621-825a-be04300e0bc3","Type":"ContainerStarted","Data":"78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59"} Dec 17 08:36:29 crc kubenswrapper[4966]: I1217 08:36:29.364352 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-glcd5" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="registry-server" probeResult="failure" output=< Dec 17 08:36:29 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:36:29 crc kubenswrapper[4966]: > Dec 17 08:36:29 crc kubenswrapper[4966]: I1217 08:36:29.381958 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-6z6kr" Dec 17 08:36:29 crc kubenswrapper[4966]: I1217 08:36:29.405705 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f76f4954c-mfghn" Dec 17 08:36:30 crc kubenswrapper[4966]: I1217 08:36:30.246120 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-4fzss" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerName="registry-server" probeResult="failure" output=< Dec 17 08:36:30 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:36:30 crc kubenswrapper[4966]: > Dec 17 08:36:30 crc kubenswrapper[4966]: I1217 08:36:30.358004 4966 generic.go:334] "Generic (PLEG): container finished" podID="55f1659f-c016-4621-825a-be04300e0bc3" containerID="78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59" exitCode=0 Dec 17 08:36:30 crc kubenswrapper[4966]: I1217 08:36:30.358062 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvd4j" event={"ID":"55f1659f-c016-4621-825a-be04300e0bc3","Type":"ContainerDied","Data":"78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59"} Dec 17 08:36:31 crc kubenswrapper[4966]: I1217 08:36:31.013439 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" Dec 17 08:36:31 crc kubenswrapper[4966]: I1217 08:36:31.765400 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-795j5" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="registry-server" probeResult="failure" output=< Dec 17 08:36:31 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:36:31 crc kubenswrapper[4966]: > Dec 17 08:36:32 crc kubenswrapper[4966]: I1217 08:36:32.372517 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvd4j" event={"ID":"55f1659f-c016-4621-825a-be04300e0bc3","Type":"ContainerStarted","Data":"335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb"} Dec 17 08:36:32 crc kubenswrapper[4966]: I1217 08:36:32.398444 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bvd4j" podStartSLOduration=4.275178629 podStartE2EDuration="7.398426489s" podCreationTimestamp="2025-12-17 08:36:25 +0000 UTC" firstStartedPulling="2025-12-17 08:36:28.323116762 +0000 UTC m=+923.868186704" lastFinishedPulling="2025-12-17 08:36:31.446364622 +0000 UTC m=+926.991434564" observedRunningTime="2025-12-17 08:36:32.395783406 +0000 UTC m=+927.940853358" watchObservedRunningTime="2025-12-17 08:36:32.398426489 +0000 UTC m=+927.943496431" Dec 17 08:36:35 crc kubenswrapper[4966]: I1217 08:36:35.533853 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" Dec 17 08:36:36 crc kubenswrapper[4966]: I1217 08:36:36.175405 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:36 crc kubenswrapper[4966]: I1217 08:36:36.175459 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:36 crc kubenswrapper[4966]: I1217 08:36:36.221378 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:36 crc kubenswrapper[4966]: I1217 08:36:36.444209 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:36 crc kubenswrapper[4966]: I1217 08:36:36.490615 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bvd4j"] Dec 17 08:36:37 crc kubenswrapper[4966]: I1217 08:36:37.832484 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.414248 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" event={"ID":"5277f562-0c42-4ca8-b8a5-aa345d02fca7","Type":"ContainerStarted","Data":"1aa1df574192c2fe78fa8d922a670111746b70970c625d2a716a4cac346d1f45"} Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.414419 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bvd4j" podUID="55f1659f-c016-4621-825a-be04300e0bc3" containerName="registry-server" containerID="cri-o://335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb" gracePeriod=2 Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.414665 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.439422 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" podStartSLOduration=3.623279084 podStartE2EDuration="1m0.439404362s" podCreationTimestamp="2025-12-17 08:35:38 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.443017625 +0000 UTC m=+876.988087567" lastFinishedPulling="2025-12-17 08:36:38.259142903 +0000 UTC m=+933.804212845" observedRunningTime="2025-12-17 08:36:38.434182969 +0000 UTC m=+933.979252911" watchObservedRunningTime="2025-12-17 08:36:38.439404362 +0000 UTC m=+933.984474304" Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.855037 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.983083 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-utilities\") pod \"55f1659f-c016-4621-825a-be04300e0bc3\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.983148 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-catalog-content\") pod \"55f1659f-c016-4621-825a-be04300e0bc3\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.983176 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k2wk\" (UniqueName: \"kubernetes.io/projected/55f1659f-c016-4621-825a-be04300e0bc3-kube-api-access-8k2wk\") pod \"55f1659f-c016-4621-825a-be04300e0bc3\" (UID: \"55f1659f-c016-4621-825a-be04300e0bc3\") " Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.983795 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-utilities" (OuterVolumeSpecName: "utilities") pod "55f1659f-c016-4621-825a-be04300e0bc3" (UID: "55f1659f-c016-4621-825a-be04300e0bc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:36:38 crc kubenswrapper[4966]: I1217 08:36:38.990408 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55f1659f-c016-4621-825a-be04300e0bc3-kube-api-access-8k2wk" (OuterVolumeSpecName: "kube-api-access-8k2wk") pod "55f1659f-c016-4621-825a-be04300e0bc3" (UID: "55f1659f-c016-4621-825a-be04300e0bc3"). InnerVolumeSpecName "kube-api-access-8k2wk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.026173 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55f1659f-c016-4621-825a-be04300e0bc3" (UID: "55f1659f-c016-4621-825a-be04300e0bc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.084174 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.084209 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55f1659f-c016-4621-825a-be04300e0bc3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.084223 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k2wk\" (UniqueName: \"kubernetes.io/projected/55f1659f-c016-4621-825a-be04300e0bc3-kube-api-access-8k2wk\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.249985 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-f458558d7-f98m5" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.256251 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.298671 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5c7cbf548f-zmxmt" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.351553 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.379278 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-glcd5" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="registry-server" probeResult="failure" output=< Dec 17 08:36:39 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:36:39 crc kubenswrapper[4966]: > Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.421084 4966 generic.go:334] "Generic (PLEG): container finished" podID="55f1659f-c016-4621-825a-be04300e0bc3" containerID="335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb" exitCode=0 Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.421139 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvd4j" event={"ID":"55f1659f-c016-4621-825a-be04300e0bc3","Type":"ContainerDied","Data":"335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb"} Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.421164 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bvd4j" event={"ID":"55f1659f-c016-4621-825a-be04300e0bc3","Type":"ContainerDied","Data":"c9982a4d35cbba486de2e655e415d7b881869c865bb25417475fe8efebd7ebb5"} Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.421179 4966 scope.go:117] "RemoveContainer" containerID="335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.421280 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bvd4j" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.423065 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" event={"ID":"861f02ab-0025-4dbb-b020-e02561b3830e","Type":"ContainerStarted","Data":"e38ff4607c4c6d85fbad6f7671abf2e61b6293c75c3f7a02d5672c22a4e186ea"} Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.440350 4966 scope.go:117] "RemoveContainer" containerID="78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.474862 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" podStartSLOduration=3.414193289 podStartE2EDuration="1m0.474846494s" podCreationTimestamp="2025-12-17 08:35:39 +0000 UTC" firstStartedPulling="2025-12-17 08:35:41.450131181 +0000 UTC m=+876.995201123" lastFinishedPulling="2025-12-17 08:36:38.510784386 +0000 UTC m=+934.055854328" observedRunningTime="2025-12-17 08:36:39.444241992 +0000 UTC m=+934.989311934" watchObservedRunningTime="2025-12-17 08:36:39.474846494 +0000 UTC m=+935.019916436" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.478991 4966 scope.go:117] "RemoveContainer" containerID="40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.480955 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bvd4j"] Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.487637 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bvd4j"] Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.525571 4966 scope.go:117] "RemoveContainer" containerID="335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb" Dec 17 08:36:39 crc kubenswrapper[4966]: E1217 08:36:39.525945 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb\": container with ID starting with 335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb not found: ID does not exist" containerID="335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.526049 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb"} err="failed to get container status \"335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb\": rpc error: code = NotFound desc = could not find container \"335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb\": container with ID starting with 335a12fe8a289df57549d61c7ac15702deb64249693a6bf744b33ebf0c5b86eb not found: ID does not exist" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.526127 4966 scope.go:117] "RemoveContainer" containerID="78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59" Dec 17 08:36:39 crc kubenswrapper[4966]: E1217 08:36:39.526424 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59\": container with ID starting with 78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59 not found: ID does not exist" containerID="78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.526449 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59"} err="failed to get container status \"78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59\": rpc error: code = NotFound desc = could not find container \"78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59\": container with ID starting with 78293520d9ce25cc54e9719f57cb3cddc87c4462f15aa374979d97d657aeab59 not found: ID does not exist" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.526464 4966 scope.go:117] "RemoveContainer" containerID="40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756" Dec 17 08:36:39 crc kubenswrapper[4966]: E1217 08:36:39.526637 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756\": container with ID starting with 40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756 not found: ID does not exist" containerID="40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.526660 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756"} err="failed to get container status \"40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756\": rpc error: code = NotFound desc = could not find container \"40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756\": container with ID starting with 40f1580e7ff9d6ff4161c9d4bd9e79b04268db832a5ff57a86d7b10d3e940756 not found: ID does not exist" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.560298 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-cr8zh" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.814769 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5c6df8f9-rcsdh" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.856640 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" Dec 17 08:36:39 crc kubenswrapper[4966]: I1217 08:36:39.858951 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-97d456b9-g6xkf" Dec 17 08:36:40 crc kubenswrapper[4966]: I1217 08:36:40.720641 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:36:40 crc kubenswrapper[4966]: I1217 08:36:40.768646 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:36:40 crc kubenswrapper[4966]: I1217 08:36:40.849425 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55f1659f-c016-4621-825a-be04300e0bc3" path="/var/lib/kubelet/pods/55f1659f-c016-4621-825a-be04300e0bc3/volumes" Dec 17 08:36:41 crc kubenswrapper[4966]: I1217 08:36:41.658919 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4fzss"] Dec 17 08:36:41 crc kubenswrapper[4966]: I1217 08:36:41.659772 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4fzss" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerName="registry-server" containerID="cri-o://f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa" gracePeriod=2 Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.068154 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.226991 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smdqc\" (UniqueName: \"kubernetes.io/projected/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-kube-api-access-smdqc\") pod \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.227155 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-utilities\") pod \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.227745 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-utilities" (OuterVolumeSpecName: "utilities") pod "5f81a852-06ec-405a-ba4e-e68b12b6e5e0" (UID: "5f81a852-06ec-405a-ba4e-e68b12b6e5e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.227865 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-catalog-content\") pod \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\" (UID: \"5f81a852-06ec-405a-ba4e-e68b12b6e5e0\") " Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.230430 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.240795 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-kube-api-access-smdqc" (OuterVolumeSpecName: "kube-api-access-smdqc") pod "5f81a852-06ec-405a-ba4e-e68b12b6e5e0" (UID: "5f81a852-06ec-405a-ba4e-e68b12b6e5e0"). InnerVolumeSpecName "kube-api-access-smdqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.248141 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f81a852-06ec-405a-ba4e-e68b12b6e5e0" (UID: "5f81a852-06ec-405a-ba4e-e68b12b6e5e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.331493 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.331535 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smdqc\" (UniqueName: \"kubernetes.io/projected/5f81a852-06ec-405a-ba4e-e68b12b6e5e0-kube-api-access-smdqc\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.451812 4966 generic.go:334] "Generic (PLEG): container finished" podID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerID="f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa" exitCode=0 Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.451950 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4fzss" event={"ID":"5f81a852-06ec-405a-ba4e-e68b12b6e5e0","Type":"ContainerDied","Data":"f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa"} Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.452011 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4fzss" event={"ID":"5f81a852-06ec-405a-ba4e-e68b12b6e5e0","Type":"ContainerDied","Data":"a58ee74eac10496c5dccac4b6479a0d2371edb5d1f34014873d62fb784c20def"} Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.452050 4966 scope.go:117] "RemoveContainer" containerID="f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.452078 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4fzss" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.471096 4966 scope.go:117] "RemoveContainer" containerID="9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.489219 4966 scope.go:117] "RemoveContainer" containerID="32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.524547 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4fzss"] Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.525625 4966 scope.go:117] "RemoveContainer" containerID="f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa" Dec 17 08:36:42 crc kubenswrapper[4966]: E1217 08:36:42.526067 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa\": container with ID starting with f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa not found: ID does not exist" containerID="f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.526099 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa"} err="failed to get container status \"f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa\": rpc error: code = NotFound desc = could not find container \"f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa\": container with ID starting with f0073f274a1ac7153d7618f83c72c5ed17bf286a7554e4fd44c96dce601f61fa not found: ID does not exist" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.526132 4966 scope.go:117] "RemoveContainer" containerID="9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5" Dec 17 08:36:42 crc kubenswrapper[4966]: E1217 08:36:42.526621 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5\": container with ID starting with 9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5 not found: ID does not exist" containerID="9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.526666 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5"} err="failed to get container status \"9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5\": rpc error: code = NotFound desc = could not find container \"9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5\": container with ID starting with 9c255721fd356a07f656d5ed3639953154a3fdd038ddc5455603d522c802c9b5 not found: ID does not exist" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.526695 4966 scope.go:117] "RemoveContainer" containerID="32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972" Dec 17 08:36:42 crc kubenswrapper[4966]: E1217 08:36:42.527002 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972\": container with ID starting with 32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972 not found: ID does not exist" containerID="32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.527026 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972"} err="failed to get container status \"32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972\": rpc error: code = NotFound desc = could not find container \"32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972\": container with ID starting with 32f32fb94555fd91c2ac5c7faca7737f1f31762963874117da55e144209f8972 not found: ID does not exist" Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.529514 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4fzss"] Dec 17 08:36:42 crc kubenswrapper[4966]: I1217 08:36:42.838694 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" path="/var/lib/kubelet/pods/5f81a852-06ec-405a-ba4e-e68b12b6e5e0/volumes" Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.065663 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-795j5"] Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.066277 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-795j5" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="registry-server" containerID="cri-o://9360b2c3dc58d21759bb96025da73cb17cd208ea4d9e5918c47555a440fccd99" gracePeriod=2 Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.470143 4966 generic.go:334] "Generic (PLEG): container finished" podID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerID="9360b2c3dc58d21759bb96025da73cb17cd208ea4d9e5918c47555a440fccd99" exitCode=0 Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.470212 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795j5" event={"ID":"057025a8-6d42-4a30-8156-b041e8c4d9aa","Type":"ContainerDied","Data":"9360b2c3dc58d21759bb96025da73cb17cd208ea4d9e5918c47555a440fccd99"} Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.470397 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-795j5" event={"ID":"057025a8-6d42-4a30-8156-b041e8c4d9aa","Type":"ContainerDied","Data":"517b2299953334b07dce6bfebdc2344e8d8354ba2b67d15f9e75081fb0845778"} Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.470409 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="517b2299953334b07dce6bfebdc2344e8d8354ba2b67d15f9e75081fb0845778" Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.476574 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.572315 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-utilities\") pod \"057025a8-6d42-4a30-8156-b041e8c4d9aa\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.572366 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-catalog-content\") pod \"057025a8-6d42-4a30-8156-b041e8c4d9aa\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.572480 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn5d5\" (UniqueName: \"kubernetes.io/projected/057025a8-6d42-4a30-8156-b041e8c4d9aa-kube-api-access-zn5d5\") pod \"057025a8-6d42-4a30-8156-b041e8c4d9aa\" (UID: \"057025a8-6d42-4a30-8156-b041e8c4d9aa\") " Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.573761 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-utilities" (OuterVolumeSpecName: "utilities") pod "057025a8-6d42-4a30-8156-b041e8c4d9aa" (UID: "057025a8-6d42-4a30-8156-b041e8c4d9aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.582077 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057025a8-6d42-4a30-8156-b041e8c4d9aa-kube-api-access-zn5d5" (OuterVolumeSpecName: "kube-api-access-zn5d5") pod "057025a8-6d42-4a30-8156-b041e8c4d9aa" (UID: "057025a8-6d42-4a30-8156-b041e8c4d9aa"). InnerVolumeSpecName "kube-api-access-zn5d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.648935 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "057025a8-6d42-4a30-8156-b041e8c4d9aa" (UID: "057025a8-6d42-4a30-8156-b041e8c4d9aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.674059 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.674087 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057025a8-6d42-4a30-8156-b041e8c4d9aa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:44 crc kubenswrapper[4966]: I1217 08:36:44.674097 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn5d5\" (UniqueName: \"kubernetes.io/projected/057025a8-6d42-4a30-8156-b041e8c4d9aa-kube-api-access-zn5d5\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:45 crc kubenswrapper[4966]: I1217 08:36:45.477596 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-795j5" Dec 17 08:36:45 crc kubenswrapper[4966]: I1217 08:36:45.497055 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-795j5"] Dec 17 08:36:45 crc kubenswrapper[4966]: I1217 08:36:45.504730 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-795j5"] Dec 17 08:36:46 crc kubenswrapper[4966]: I1217 08:36:46.841862 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" path="/var/lib/kubelet/pods/057025a8-6d42-4a30-8156-b041e8c4d9aa/volumes" Dec 17 08:36:48 crc kubenswrapper[4966]: I1217 08:36:48.381395 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:48 crc kubenswrapper[4966]: I1217 08:36:48.477011 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:49 crc kubenswrapper[4966]: I1217 08:36:49.260221 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-glcd5"] Dec 17 08:36:49 crc kubenswrapper[4966]: I1217 08:36:49.506496 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-glcd5" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="registry-server" containerID="cri-o://64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099" gracePeriod=2 Dec 17 08:36:49 crc kubenswrapper[4966]: I1217 08:36:49.796935 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-756ccf86c7-ksgt7" Dec 17 08:36:49 crc kubenswrapper[4966]: I1217 08:36:49.861353 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-55f78b7c4c-z9pzc" Dec 17 08:36:49 crc kubenswrapper[4966]: I1217 08:36:49.886128 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.063940 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-utilities\") pod \"8d2c4ece-23cd-4941-9a72-9958ee533c09\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.064095 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkrtf\" (UniqueName: \"kubernetes.io/projected/8d2c4ece-23cd-4941-9a72-9958ee533c09-kube-api-access-bkrtf\") pod \"8d2c4ece-23cd-4941-9a72-9958ee533c09\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.064119 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-catalog-content\") pod \"8d2c4ece-23cd-4941-9a72-9958ee533c09\" (UID: \"8d2c4ece-23cd-4941-9a72-9958ee533c09\") " Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.065322 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-utilities" (OuterVolumeSpecName: "utilities") pod "8d2c4ece-23cd-4941-9a72-9958ee533c09" (UID: "8d2c4ece-23cd-4941-9a72-9958ee533c09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.069919 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2c4ece-23cd-4941-9a72-9958ee533c09-kube-api-access-bkrtf" (OuterVolumeSpecName: "kube-api-access-bkrtf") pod "8d2c4ece-23cd-4941-9a72-9958ee533c09" (UID: "8d2c4ece-23cd-4941-9a72-9958ee533c09"). InnerVolumeSpecName "kube-api-access-bkrtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.165377 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.165408 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkrtf\" (UniqueName: \"kubernetes.io/projected/8d2c4ece-23cd-4941-9a72-9958ee533c09-kube-api-access-bkrtf\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.186354 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d2c4ece-23cd-4941-9a72-9958ee533c09" (UID: "8d2c4ece-23cd-4941-9a72-9958ee533c09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.266824 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2c4ece-23cd-4941-9a72-9958ee533c09-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.519403 4966 generic.go:334] "Generic (PLEG): container finished" podID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerID="64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099" exitCode=0 Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.519447 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-glcd5" event={"ID":"8d2c4ece-23cd-4941-9a72-9958ee533c09","Type":"ContainerDied","Data":"64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099"} Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.519485 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-glcd5" event={"ID":"8d2c4ece-23cd-4941-9a72-9958ee533c09","Type":"ContainerDied","Data":"e7082b97031349cd4983b0ba83f5916681f18fab25677754d6206b847a1a5018"} Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.519505 4966 scope.go:117] "RemoveContainer" containerID="64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.519548 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-glcd5" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.548913 4966 scope.go:117] "RemoveContainer" containerID="4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.606216 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-glcd5"] Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.607664 4966 scope.go:117] "RemoveContainer" containerID="d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.625491 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-glcd5"] Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.644124 4966 scope.go:117] "RemoveContainer" containerID="64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099" Dec 17 08:36:50 crc kubenswrapper[4966]: E1217 08:36:50.646750 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099\": container with ID starting with 64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099 not found: ID does not exist" containerID="64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.646792 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099"} err="failed to get container status \"64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099\": rpc error: code = NotFound desc = could not find container \"64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099\": container with ID starting with 64075d934dda5ab2b5c1b46b67dcd643439ce9639dc9cdea942e1f824247e099 not found: ID does not exist" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.646817 4966 scope.go:117] "RemoveContainer" containerID="4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c" Dec 17 08:36:50 crc kubenswrapper[4966]: E1217 08:36:50.647243 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c\": container with ID starting with 4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c not found: ID does not exist" containerID="4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.647278 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c"} err="failed to get container status \"4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c\": rpc error: code = NotFound desc = could not find container \"4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c\": container with ID starting with 4236d4266763a9f7275e0f341b141ea5716183542834b6b5c4acdb2106c67f8c not found: ID does not exist" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.647294 4966 scope.go:117] "RemoveContainer" containerID="d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194" Dec 17 08:36:50 crc kubenswrapper[4966]: E1217 08:36:50.647591 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194\": container with ID starting with d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194 not found: ID does not exist" containerID="d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.647621 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194"} err="failed to get container status \"d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194\": rpc error: code = NotFound desc = could not find container \"d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194\": container with ID starting with d9d077824c24f4596f97f8a60af5281d608a1f1bc5fe6ca633f4eba3158fd194 not found: ID does not exist" Dec 17 08:36:50 crc kubenswrapper[4966]: I1217 08:36:50.840496 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" path="/var/lib/kubelet/pods/8d2c4ece-23cd-4941-9a72-9958ee533c09/volumes" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.925134 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b75798c-q5gqg"] Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.925843 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="extract-utilities" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.925855 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="extract-utilities" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.925889 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerName="extract-content" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.925903 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerName="extract-content" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.925918 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerName="extract-utilities" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.925925 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerName="extract-utilities" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.925935 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.925940 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.925960 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55f1659f-c016-4621-825a-be04300e0bc3" containerName="extract-utilities" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.925967 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="55f1659f-c016-4621-825a-be04300e0bc3" containerName="extract-utilities" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.925974 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.925979 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.926000 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="extract-content" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926006 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="extract-content" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.926014 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="extract-utilities" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926020 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="extract-utilities" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.926034 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55f1659f-c016-4621-825a-be04300e0bc3" containerName="extract-content" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926042 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="55f1659f-c016-4621-825a-be04300e0bc3" containerName="extract-content" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.926055 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55f1659f-c016-4621-825a-be04300e0bc3" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926060 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="55f1659f-c016-4621-825a-be04300e0bc3" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.926070 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926077 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: E1217 08:37:04.926085 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="extract-content" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926090 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="extract-content" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926227 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="55f1659f-c016-4621-825a-be04300e0bc3" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926241 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f81a852-06ec-405a-ba4e-e68b12b6e5e0" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926249 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="057025a8-6d42-4a30-8156-b041e8c4d9aa" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926261 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2c4ece-23cd-4941-9a72-9958ee533c09" containerName="registry-server" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.926943 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.933934 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-pzt6k" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.934153 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.934278 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.934451 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 17 08:37:04 crc kubenswrapper[4966]: I1217 08:37:04.961994 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b75798c-q5gqg"] Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.036081 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64d8cd5c57-stcsp"] Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.037142 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.042247 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.065399 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d8cd5c57-stcsp"] Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.082830 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618c6158-f35f-4a0b-a1ba-fdee7eeed869-config\") pod \"dnsmasq-dns-84b75798c-q5gqg\" (UID: \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\") " pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.082924 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnq9s\" (UniqueName: \"kubernetes.io/projected/618c6158-f35f-4a0b-a1ba-fdee7eeed869-kube-api-access-jnq9s\") pod \"dnsmasq-dns-84b75798c-q5gqg\" (UID: \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\") " pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.184816 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618c6158-f35f-4a0b-a1ba-fdee7eeed869-config\") pod \"dnsmasq-dns-84b75798c-q5gqg\" (UID: \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\") " pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.185246 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnq9s\" (UniqueName: \"kubernetes.io/projected/618c6158-f35f-4a0b-a1ba-fdee7eeed869-kube-api-access-jnq9s\") pod \"dnsmasq-dns-84b75798c-q5gqg\" (UID: \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\") " pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.185307 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns9m6\" (UniqueName: \"kubernetes.io/projected/ab331b3a-89a2-4d30-8c93-eec6b456d9be-kube-api-access-ns9m6\") pod \"dnsmasq-dns-64d8cd5c57-stcsp\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.185345 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-config\") pod \"dnsmasq-dns-64d8cd5c57-stcsp\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.185426 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-dns-svc\") pod \"dnsmasq-dns-64d8cd5c57-stcsp\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.185821 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618c6158-f35f-4a0b-a1ba-fdee7eeed869-config\") pod \"dnsmasq-dns-84b75798c-q5gqg\" (UID: \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\") " pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.216810 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnq9s\" (UniqueName: \"kubernetes.io/projected/618c6158-f35f-4a0b-a1ba-fdee7eeed869-kube-api-access-jnq9s\") pod \"dnsmasq-dns-84b75798c-q5gqg\" (UID: \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\") " pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.251594 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.286207 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-dns-svc\") pod \"dnsmasq-dns-64d8cd5c57-stcsp\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.286322 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns9m6\" (UniqueName: \"kubernetes.io/projected/ab331b3a-89a2-4d30-8c93-eec6b456d9be-kube-api-access-ns9m6\") pod \"dnsmasq-dns-64d8cd5c57-stcsp\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.286346 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-config\") pod \"dnsmasq-dns-64d8cd5c57-stcsp\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.287148 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-dns-svc\") pod \"dnsmasq-dns-64d8cd5c57-stcsp\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.287285 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-config\") pod \"dnsmasq-dns-64d8cd5c57-stcsp\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.304966 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns9m6\" (UniqueName: \"kubernetes.io/projected/ab331b3a-89a2-4d30-8c93-eec6b456d9be-kube-api-access-ns9m6\") pod \"dnsmasq-dns-64d8cd5c57-stcsp\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.363925 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.694074 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b75798c-q5gqg"] Dec 17 08:37:05 crc kubenswrapper[4966]: W1217 08:37:05.826349 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab331b3a_89a2_4d30_8c93_eec6b456d9be.slice/crio-c0ccdab5a61fb118adbca23c1cace00a29969c42aa3972c3983e73e9da486dff WatchSource:0}: Error finding container c0ccdab5a61fb118adbca23c1cace00a29969c42aa3972c3983e73e9da486dff: Status 404 returned error can't find the container with id c0ccdab5a61fb118adbca23c1cace00a29969c42aa3972c3983e73e9da486dff Dec 17 08:37:05 crc kubenswrapper[4966]: I1217 08:37:05.832023 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64d8cd5c57-stcsp"] Dec 17 08:37:06 crc kubenswrapper[4966]: I1217 08:37:06.645765 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b75798c-q5gqg" event={"ID":"618c6158-f35f-4a0b-a1ba-fdee7eeed869","Type":"ContainerStarted","Data":"0935f0085565b181835a7d8472a49012eb1008b1b5307adfdda1dd5b561b533c"} Dec 17 08:37:06 crc kubenswrapper[4966]: I1217 08:37:06.648015 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" event={"ID":"ab331b3a-89a2-4d30-8c93-eec6b456d9be","Type":"ContainerStarted","Data":"c0ccdab5a61fb118adbca23c1cace00a29969c42aa3972c3983e73e9da486dff"} Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.054766 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b75798c-q5gqg"] Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.080555 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78d6cd766f-w5jdf"] Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.089228 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.107516 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78d6cd766f-w5jdf"] Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.239990 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-dns-svc\") pod \"dnsmasq-dns-78d6cd766f-w5jdf\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.240063 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-config\") pod \"dnsmasq-dns-78d6cd766f-w5jdf\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.240118 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x99b9\" (UniqueName: \"kubernetes.io/projected/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-kube-api-access-x99b9\") pod \"dnsmasq-dns-78d6cd766f-w5jdf\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.341757 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-dns-svc\") pod \"dnsmasq-dns-78d6cd766f-w5jdf\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.342052 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-config\") pod \"dnsmasq-dns-78d6cd766f-w5jdf\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.342109 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x99b9\" (UniqueName: \"kubernetes.io/projected/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-kube-api-access-x99b9\") pod \"dnsmasq-dns-78d6cd766f-w5jdf\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.342918 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-config\") pod \"dnsmasq-dns-78d6cd766f-w5jdf\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.342968 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-dns-svc\") pod \"dnsmasq-dns-78d6cd766f-w5jdf\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.387727 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x99b9\" (UniqueName: \"kubernetes.io/projected/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-kube-api-access-x99b9\") pod \"dnsmasq-dns-78d6cd766f-w5jdf\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.393336 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d8cd5c57-stcsp"] Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.415266 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.425805 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-647ff6c755-tzk8k"] Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.427084 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.438859 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-647ff6c755-tzk8k"] Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.547328 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-dns-svc\") pod \"dnsmasq-dns-647ff6c755-tzk8k\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.547428 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-config\") pod \"dnsmasq-dns-647ff6c755-tzk8k\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.547462 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snd6c\" (UniqueName: \"kubernetes.io/projected/3e81f86f-d072-4057-807d-81d017f072f9-kube-api-access-snd6c\") pod \"dnsmasq-dns-647ff6c755-tzk8k\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.649987 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-config\") pod \"dnsmasq-dns-647ff6c755-tzk8k\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.650248 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snd6c\" (UniqueName: \"kubernetes.io/projected/3e81f86f-d072-4057-807d-81d017f072f9-kube-api-access-snd6c\") pod \"dnsmasq-dns-647ff6c755-tzk8k\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.650320 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-dns-svc\") pod \"dnsmasq-dns-647ff6c755-tzk8k\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.651006 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-dns-svc\") pod \"dnsmasq-dns-647ff6c755-tzk8k\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.653743 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-config\") pod \"dnsmasq-dns-647ff6c755-tzk8k\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.676229 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snd6c\" (UniqueName: \"kubernetes.io/projected/3e81f86f-d072-4057-807d-81d017f072f9-kube-api-access-snd6c\") pod \"dnsmasq-dns-647ff6c755-tzk8k\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:08 crc kubenswrapper[4966]: I1217 08:37:08.755478 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.019535 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78d6cd766f-w5jdf"] Dec 17 08:37:09 crc kubenswrapper[4966]: W1217 08:37:09.034910 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3ae48f4_843d_4c46_9fd7_5f27f10863d1.slice/crio-4be211082464b8bb8c84e8d40b8ae6e5e4800d276ef10d4abb40d1c03ebf6b99 WatchSource:0}: Error finding container 4be211082464b8bb8c84e8d40b8ae6e5e4800d276ef10d4abb40d1c03ebf6b99: Status 404 returned error can't find the container with id 4be211082464b8bb8c84e8d40b8ae6e5e4800d276ef10d4abb40d1c03ebf6b99 Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.180853 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.182934 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.185413 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.186826 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.187162 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.187182 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.187435 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ft6gd" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.188523 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.189464 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.191893 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.255099 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-647ff6c755-tzk8k"] Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.262857 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.262920 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bd01548-3d59-4472-b8ef-a899b2cedee8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.262945 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.262966 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.262987 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.263046 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.263187 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.263218 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.263322 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.263353 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rbpr\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-kube-api-access-9rbpr\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.263387 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bd01548-3d59-4472-b8ef-a899b2cedee8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.364900 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.364948 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.364970 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.364999 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.365075 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rbpr\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-kube-api-access-9rbpr\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.365104 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bd01548-3d59-4472-b8ef-a899b2cedee8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.365131 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.365148 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bd01548-3d59-4472-b8ef-a899b2cedee8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.365165 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.365189 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.365220 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.366044 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.366799 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.367175 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.368150 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.368934 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.370742 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.372561 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.373024 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bd01548-3d59-4472-b8ef-a899b2cedee8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.373395 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bd01548-3d59-4472-b8ef-a899b2cedee8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.383628 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rbpr\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-kube-api-access-9rbpr\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.393381 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.425328 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.534833 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.568762 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.569963 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.571774 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.572000 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.573784 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.573965 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.574046 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.574122 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.579470 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vj9tw" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.591151 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.668922 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.668972 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccbzf\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-kube-api-access-ccbzf\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.669000 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.669050 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.669067 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.669091 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.669114 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.669130 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.669161 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.669179 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-config-data\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.669192 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.691829 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" event={"ID":"e3ae48f4-843d-4c46-9fd7-5f27f10863d1","Type":"ContainerStarted","Data":"4be211082464b8bb8c84e8d40b8ae6e5e4800d276ef10d4abb40d1c03ebf6b99"} Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770244 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770301 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770325 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770362 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770384 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-config-data\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770401 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770422 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770446 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccbzf\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-kube-api-access-ccbzf\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770466 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770515 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770532 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.770794 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.771400 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.773091 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-config-data\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.773679 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.775511 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.786107 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.787357 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccbzf\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-kube-api-access-ccbzf\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.789670 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.790251 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.791500 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.792264 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.807523 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " pod="openstack/rabbitmq-server-0" Dec 17 08:37:09 crc kubenswrapper[4966]: I1217 08:37:09.905807 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.283067 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.285001 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.298803 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.299054 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.299162 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-d6hrp" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.306518 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.311894 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.337551 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.402224 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b562eab8-f0a5-4de7-a251-367555da04bf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.402271 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b562eab8-f0a5-4de7-a251-367555da04bf-kolla-config\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.402298 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b562eab8-f0a5-4de7-a251-367555da04bf-config-data-default\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.402340 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b562eab8-f0a5-4de7-a251-367555da04bf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.402353 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b562eab8-f0a5-4de7-a251-367555da04bf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.402376 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.402391 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq27m\" (UniqueName: \"kubernetes.io/projected/b562eab8-f0a5-4de7-a251-367555da04bf-kube-api-access-pq27m\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.402409 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b562eab8-f0a5-4de7-a251-367555da04bf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.503821 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b562eab8-f0a5-4de7-a251-367555da04bf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.503864 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b562eab8-f0a5-4de7-a251-367555da04bf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.503913 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.503930 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq27m\" (UniqueName: \"kubernetes.io/projected/b562eab8-f0a5-4de7-a251-367555da04bf-kube-api-access-pq27m\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.503951 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b562eab8-f0a5-4de7-a251-367555da04bf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.504361 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b562eab8-f0a5-4de7-a251-367555da04bf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.505420 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.506468 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b562eab8-f0a5-4de7-a251-367555da04bf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.506558 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b562eab8-f0a5-4de7-a251-367555da04bf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.506589 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b562eab8-f0a5-4de7-a251-367555da04bf-kolla-config\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.506612 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b562eab8-f0a5-4de7-a251-367555da04bf-config-data-default\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.507303 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b562eab8-f0a5-4de7-a251-367555da04bf-config-data-default\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.507937 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b562eab8-f0a5-4de7-a251-367555da04bf-kolla-config\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.520723 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b562eab8-f0a5-4de7-a251-367555da04bf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.521412 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b562eab8-f0a5-4de7-a251-367555da04bf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.524193 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq27m\" (UniqueName: \"kubernetes.io/projected/b562eab8-f0a5-4de7-a251-367555da04bf-kube-api-access-pq27m\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.537865 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"b562eab8-f0a5-4de7-a251-367555da04bf\") " pod="openstack/openstack-galera-0" Dec 17 08:37:11 crc kubenswrapper[4966]: I1217 08:37:11.632763 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.447483 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.450093 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.460108 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.460611 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-pgfcg" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.460772 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.460820 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.464211 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.527264 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/83bca014-fd3a-430d-b532-77c1429a3728-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.527323 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbzct\" (UniqueName: \"kubernetes.io/projected/83bca014-fd3a-430d-b532-77c1429a3728-kube-api-access-gbzct\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.527369 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/83bca014-fd3a-430d-b532-77c1429a3728-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.527410 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.527430 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/83bca014-fd3a-430d-b532-77c1429a3728-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.527450 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bca014-fd3a-430d-b532-77c1429a3728-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.527477 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/83bca014-fd3a-430d-b532-77c1429a3728-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.527512 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83bca014-fd3a-430d-b532-77c1429a3728-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.634652 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/83bca014-fd3a-430d-b532-77c1429a3728-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.634710 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.634767 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/83bca014-fd3a-430d-b532-77c1429a3728-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.635088 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.635164 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bca014-fd3a-430d-b532-77c1429a3728-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.637096 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/83bca014-fd3a-430d-b532-77c1429a3728-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.637562 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/83bca014-fd3a-430d-b532-77c1429a3728-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.636031 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/83bca014-fd3a-430d-b532-77c1429a3728-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.636741 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bca014-fd3a-430d-b532-77c1429a3728-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.644958 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83bca014-fd3a-430d-b532-77c1429a3728-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.645029 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/83bca014-fd3a-430d-b532-77c1429a3728-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.645081 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbzct\" (UniqueName: \"kubernetes.io/projected/83bca014-fd3a-430d-b532-77c1429a3728-kube-api-access-gbzct\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.645781 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/83bca014-fd3a-430d-b532-77c1429a3728-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.650486 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/83bca014-fd3a-430d-b532-77c1429a3728-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.650507 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83bca014-fd3a-430d-b532-77c1429a3728-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.664043 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.684315 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbzct\" (UniqueName: \"kubernetes.io/projected/83bca014-fd3a-430d-b532-77c1429a3728-kube-api-access-gbzct\") pod \"openstack-cell1-galera-0\" (UID: \"83bca014-fd3a-430d-b532-77c1429a3728\") " pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.782514 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.794927 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.795845 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.808937 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.809103 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-bbbtj" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.809843 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.814604 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.849116 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d362072-2095-4d41-8ab5-7681c040faf6-config-data\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.849188 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d362072-2095-4d41-8ab5-7681c040faf6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.849278 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d362072-2095-4d41-8ab5-7681c040faf6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.849634 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6d362072-2095-4d41-8ab5-7681c040faf6-kolla-config\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.849691 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs2v8\" (UniqueName: \"kubernetes.io/projected/6d362072-2095-4d41-8ab5-7681c040faf6-kube-api-access-bs2v8\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.950945 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d362072-2095-4d41-8ab5-7681c040faf6-config-data\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.951001 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d362072-2095-4d41-8ab5-7681c040faf6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.951045 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d362072-2095-4d41-8ab5-7681c040faf6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.951145 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6d362072-2095-4d41-8ab5-7681c040faf6-kolla-config\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.951174 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs2v8\" (UniqueName: \"kubernetes.io/projected/6d362072-2095-4d41-8ab5-7681c040faf6-kube-api-access-bs2v8\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.952542 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6d362072-2095-4d41-8ab5-7681c040faf6-kolla-config\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.953504 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d362072-2095-4d41-8ab5-7681c040faf6-config-data\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.961363 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d362072-2095-4d41-8ab5-7681c040faf6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.964219 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d362072-2095-4d41-8ab5-7681c040faf6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:12 crc kubenswrapper[4966]: I1217 08:37:12.986334 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs2v8\" (UniqueName: \"kubernetes.io/projected/6d362072-2095-4d41-8ab5-7681c040faf6-kube-api-access-bs2v8\") pod \"memcached-0\" (UID: \"6d362072-2095-4d41-8ab5-7681c040faf6\") " pod="openstack/memcached-0" Dec 17 08:37:13 crc kubenswrapper[4966]: I1217 08:37:13.124180 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 17 08:37:14 crc kubenswrapper[4966]: I1217 08:37:14.989683 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 17 08:37:14 crc kubenswrapper[4966]: I1217 08:37:14.991213 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 17 08:37:14 crc kubenswrapper[4966]: I1217 08:37:14.998945 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-fcf6r" Dec 17 08:37:15 crc kubenswrapper[4966]: I1217 08:37:15.011013 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 17 08:37:15 crc kubenswrapper[4966]: I1217 08:37:15.123179 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2nbd\" (UniqueName: \"kubernetes.io/projected/bdca14dc-79c4-4e52-a748-9977656f7821-kube-api-access-t2nbd\") pod \"kube-state-metrics-0\" (UID: \"bdca14dc-79c4-4e52-a748-9977656f7821\") " pod="openstack/kube-state-metrics-0" Dec 17 08:37:15 crc kubenswrapper[4966]: I1217 08:37:15.224221 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2nbd\" (UniqueName: \"kubernetes.io/projected/bdca14dc-79c4-4e52-a748-9977656f7821-kube-api-access-t2nbd\") pod \"kube-state-metrics-0\" (UID: \"bdca14dc-79c4-4e52-a748-9977656f7821\") " pod="openstack/kube-state-metrics-0" Dec 17 08:37:15 crc kubenswrapper[4966]: I1217 08:37:15.240573 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2nbd\" (UniqueName: \"kubernetes.io/projected/bdca14dc-79c4-4e52-a748-9977656f7821-kube-api-access-t2nbd\") pod \"kube-state-metrics-0\" (UID: \"bdca14dc-79c4-4e52-a748-9977656f7821\") " pod="openstack/kube-state-metrics-0" Dec 17 08:37:15 crc kubenswrapper[4966]: I1217 08:37:15.309119 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 17 08:37:15 crc kubenswrapper[4966]: I1217 08:37:15.740603 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" event={"ID":"3e81f86f-d072-4057-807d-81d017f072f9","Type":"ContainerStarted","Data":"29b22ac1b59ac0d9501c81785144d921fcb3a4de838e9c9e85784f672b25c62a"} Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.333053 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cbln8"] Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.334379 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.341175 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cbln8"] Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.356391 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-vndpq" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.357696 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.357906 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.366048 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-f6rhs"] Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.367579 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.406977 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-f6rhs"] Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488497 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-scripts\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488558 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3cb36f72-ce5e-4490-9009-7fe4443152b5-scripts\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488586 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-var-lib\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488615 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsvht\" (UniqueName: \"kubernetes.io/projected/3cb36f72-ce5e-4490-9009-7fe4443152b5-kube-api-access-gsvht\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488638 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-var-run\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488660 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3cb36f72-ce5e-4490-9009-7fe4443152b5-var-log-ovn\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488686 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cb36f72-ce5e-4490-9009-7fe4443152b5-var-run-ovn\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488701 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cb36f72-ce5e-4490-9009-7fe4443152b5-ovn-controller-tls-certs\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488730 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3cb36f72-ce5e-4490-9009-7fe4443152b5-var-run\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488780 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-var-log\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488815 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgv64\" (UniqueName: \"kubernetes.io/projected/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-kube-api-access-mgv64\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488842 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb36f72-ce5e-4490-9009-7fe4443152b5-combined-ca-bundle\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.488997 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-etc-ovs\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.589956 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgv64\" (UniqueName: \"kubernetes.io/projected/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-kube-api-access-mgv64\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.590024 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb36f72-ce5e-4490-9009-7fe4443152b5-combined-ca-bundle\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.590722 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-etc-ovs\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.590065 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-etc-ovs\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591093 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-scripts\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591143 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3cb36f72-ce5e-4490-9009-7fe4443152b5-scripts\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591192 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-var-lib\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591242 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsvht\" (UniqueName: \"kubernetes.io/projected/3cb36f72-ce5e-4490-9009-7fe4443152b5-kube-api-access-gsvht\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591286 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-var-run\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591324 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3cb36f72-ce5e-4490-9009-7fe4443152b5-var-log-ovn\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591399 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cb36f72-ce5e-4490-9009-7fe4443152b5-var-run-ovn\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591417 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cb36f72-ce5e-4490-9009-7fe4443152b5-ovn-controller-tls-certs\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591505 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3cb36f72-ce5e-4490-9009-7fe4443152b5-var-run\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591550 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-var-log\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591841 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cb36f72-ce5e-4490-9009-7fe4443152b5-var-run-ovn\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.591891 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-var-log\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.592003 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-var-run\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.592119 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/3cb36f72-ce5e-4490-9009-7fe4443152b5-var-log-ovn\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.592180 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3cb36f72-ce5e-4490-9009-7fe4443152b5-var-run\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.592330 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-var-lib\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.594522 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-scripts\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.596018 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3cb36f72-ce5e-4490-9009-7fe4443152b5-scripts\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.596584 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cb36f72-ce5e-4490-9009-7fe4443152b5-ovn-controller-tls-certs\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.608928 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgv64\" (UniqueName: \"kubernetes.io/projected/30d8ef9a-31d4-4d75-91d2-cbf741bcf41f-kube-api-access-mgv64\") pod \"ovn-controller-ovs-f6rhs\" (UID: \"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f\") " pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.611001 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb36f72-ce5e-4490-9009-7fe4443152b5-combined-ca-bundle\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.613663 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsvht\" (UniqueName: \"kubernetes.io/projected/3cb36f72-ce5e-4490-9009-7fe4443152b5-kube-api-access-gsvht\") pod \"ovn-controller-cbln8\" (UID: \"3cb36f72-ce5e-4490-9009-7fe4443152b5\") " pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.667898 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8" Dec 17 08:37:18 crc kubenswrapper[4966]: I1217 08:37:18.703824 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.243464 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.245117 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.252483 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.252807 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-d7mrd" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.252979 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.253111 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.253815 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.258179 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.404490 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0279d039-e449-4958-aebd-2d619653460d-config\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.404547 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0279d039-e449-4958-aebd-2d619653460d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.404638 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0279d039-e449-4958-aebd-2d619653460d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.404722 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0279d039-e449-4958-aebd-2d619653460d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.404780 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.404944 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0279d039-e449-4958-aebd-2d619653460d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.405074 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzqw6\" (UniqueName: \"kubernetes.io/projected/0279d039-e449-4958-aebd-2d619653460d-kube-api-access-kzqw6\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.405134 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0279d039-e449-4958-aebd-2d619653460d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.506654 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzqw6\" (UniqueName: \"kubernetes.io/projected/0279d039-e449-4958-aebd-2d619653460d-kube-api-access-kzqw6\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.506712 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0279d039-e449-4958-aebd-2d619653460d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.506789 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0279d039-e449-4958-aebd-2d619653460d-config\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.506817 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0279d039-e449-4958-aebd-2d619653460d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.506863 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0279d039-e449-4958-aebd-2d619653460d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.506932 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0279d039-e449-4958-aebd-2d619653460d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.506972 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.507027 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0279d039-e449-4958-aebd-2d619653460d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.507921 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.508189 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0279d039-e449-4958-aebd-2d619653460d-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.509204 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0279d039-e449-4958-aebd-2d619653460d-config\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.510031 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0279d039-e449-4958-aebd-2d619653460d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.515334 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0279d039-e449-4958-aebd-2d619653460d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.519888 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0279d039-e449-4958-aebd-2d619653460d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.524448 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzqw6\" (UniqueName: \"kubernetes.io/projected/0279d039-e449-4958-aebd-2d619653460d-kube-api-access-kzqw6\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.532907 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0279d039-e449-4958-aebd-2d619653460d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.562575 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0279d039-e449-4958-aebd-2d619653460d\") " pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:19 crc kubenswrapper[4966]: I1217 08:37:19.601120 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.225629 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.227316 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.229459 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-68fgx" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.229536 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.229610 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.229698 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.233548 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.421986 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.422037 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7042a58-a47b-4eed-b4d6-625a3f57fe94-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.422058 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7042a58-a47b-4eed-b4d6-625a3f57fe94-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.422108 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7042a58-a47b-4eed-b4d6-625a3f57fe94-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.422128 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpmsc\" (UniqueName: \"kubernetes.io/projected/d7042a58-a47b-4eed-b4d6-625a3f57fe94-kube-api-access-rpmsc\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.422150 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7042a58-a47b-4eed-b4d6-625a3f57fe94-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.422189 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7042a58-a47b-4eed-b4d6-625a3f57fe94-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.422206 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7042a58-a47b-4eed-b4d6-625a3f57fe94-config\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.524133 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.524202 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7042a58-a47b-4eed-b4d6-625a3f57fe94-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.524224 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7042a58-a47b-4eed-b4d6-625a3f57fe94-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.524273 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7042a58-a47b-4eed-b4d6-625a3f57fe94-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.524293 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpmsc\" (UniqueName: \"kubernetes.io/projected/d7042a58-a47b-4eed-b4d6-625a3f57fe94-kube-api-access-rpmsc\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.524315 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7042a58-a47b-4eed-b4d6-625a3f57fe94-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.524359 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7042a58-a47b-4eed-b4d6-625a3f57fe94-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.524376 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7042a58-a47b-4eed-b4d6-625a3f57fe94-config\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.525101 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7042a58-a47b-4eed-b4d6-625a3f57fe94-config\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.525411 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.527907 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7042a58-a47b-4eed-b4d6-625a3f57fe94-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.528066 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7042a58-a47b-4eed-b4d6-625a3f57fe94-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.532090 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7042a58-a47b-4eed-b4d6-625a3f57fe94-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.532472 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7042a58-a47b-4eed-b4d6-625a3f57fe94-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.538999 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7042a58-a47b-4eed-b4d6-625a3f57fe94-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.544638 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpmsc\" (UniqueName: \"kubernetes.io/projected/d7042a58-a47b-4eed-b4d6-625a3f57fe94-kube-api-access-rpmsc\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.550309 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d7042a58-a47b-4eed-b4d6-625a3f57fe94\") " pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:22 crc kubenswrapper[4966]: I1217 08:37:22.591217 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.236051 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.236509 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.236631 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x99b9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78d6cd766f-w5jdf_openstack(e3ae48f4-843d-4c46-9fd7-5f27f10863d1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.238092 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" podUID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.238514 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.238538 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.238764 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ns9m6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-64d8cd5c57-stcsp_openstack(ab331b3a-89a2-4d30-8c93-eec6b456d9be): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.241978 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" podUID="ab331b3a-89a2-4d30-8c93-eec6b456d9be" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.261767 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.261817 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.261950 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-neutron-server:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jnq9s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-84b75798c-q5gqg_openstack(618c6158-f35f-4a0b-a1ba-fdee7eeed869): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:37:29 crc kubenswrapper[4966]: E1217 08:37:29.263748 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-84b75798c-q5gqg" podUID="618c6158-f35f-4a0b-a1ba-fdee7eeed869" Dec 17 08:37:29 crc kubenswrapper[4966]: I1217 08:37:29.788691 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 17 08:37:29 crc kubenswrapper[4966]: I1217 08:37:29.871518 4966 generic.go:334] "Generic (PLEG): container finished" podID="3e81f86f-d072-4057-807d-81d017f072f9" containerID="0ab9548d146e4a17c461faa032ae051729fabb30139069121427c5f6dc8f338a" exitCode=0 Dec 17 08:37:29 crc kubenswrapper[4966]: I1217 08:37:29.871795 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" event={"ID":"3e81f86f-d072-4057-807d-81d017f072f9","Type":"ContainerDied","Data":"0ab9548d146e4a17c461faa032ae051729fabb30139069121427c5f6dc8f338a"} Dec 17 08:37:29 crc kubenswrapper[4966]: I1217 08:37:29.874378 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1bd01548-3d59-4472-b8ef-a899b2cedee8","Type":"ContainerStarted","Data":"20b0f6f21e431555cb4d03aedbaf0f13fdd6ed875e549f3b4c4630c4cb5a49ba"} Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.256557 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.291670 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.311662 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 17 08:37:30 crc kubenswrapper[4966]: W1217 08:37:30.311712 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d362072_2095_4d41_8ab5_7681c040faf6.slice/crio-9a0e34856fdff0066c6f34bd0e1e895a7987657471192c671fef7ec03f0acc12 WatchSource:0}: Error finding container 9a0e34856fdff0066c6f34bd0e1e895a7987657471192c671fef7ec03f0acc12: Status 404 returned error can't find the container with id 9a0e34856fdff0066c6f34bd0e1e895a7987657471192c671fef7ec03f0acc12 Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.319768 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.338246 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.542989 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cbln8"] Dec 17 08:37:30 crc kubenswrapper[4966]: W1217 08:37:30.551474 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cb36f72_ce5e_4490_9009_7fe4443152b5.slice/crio-5d95e596ffe1dd62462017dc65266c898b889d8b9ad2037cac9dd9ce0773a43e WatchSource:0}: Error finding container 5d95e596ffe1dd62462017dc65266c898b889d8b9ad2037cac9dd9ce0773a43e: Status 404 returned error can't find the container with id 5d95e596ffe1dd62462017dc65266c898b889d8b9ad2037cac9dd9ce0773a43e Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.609003 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.615161 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.773583 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns9m6\" (UniqueName: \"kubernetes.io/projected/ab331b3a-89a2-4d30-8c93-eec6b456d9be-kube-api-access-ns9m6\") pod \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.774865 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-config\") pod \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.775055 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618c6158-f35f-4a0b-a1ba-fdee7eeed869-config\") pod \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\" (UID: \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\") " Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.775149 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-dns-svc\") pod \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\" (UID: \"ab331b3a-89a2-4d30-8c93-eec6b456d9be\") " Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.775260 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnq9s\" (UniqueName: \"kubernetes.io/projected/618c6158-f35f-4a0b-a1ba-fdee7eeed869-kube-api-access-jnq9s\") pod \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\" (UID: \"618c6158-f35f-4a0b-a1ba-fdee7eeed869\") " Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.777371 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/618c6158-f35f-4a0b-a1ba-fdee7eeed869-config" (OuterVolumeSpecName: "config") pod "618c6158-f35f-4a0b-a1ba-fdee7eeed869" (UID: "618c6158-f35f-4a0b-a1ba-fdee7eeed869"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.777682 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ab331b3a-89a2-4d30-8c93-eec6b456d9be" (UID: "ab331b3a-89a2-4d30-8c93-eec6b456d9be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.778209 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-config" (OuterVolumeSpecName: "config") pod "ab331b3a-89a2-4d30-8c93-eec6b456d9be" (UID: "ab331b3a-89a2-4d30-8c93-eec6b456d9be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.781605 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/618c6158-f35f-4a0b-a1ba-fdee7eeed869-kube-api-access-jnq9s" (OuterVolumeSpecName: "kube-api-access-jnq9s") pod "618c6158-f35f-4a0b-a1ba-fdee7eeed869" (UID: "618c6158-f35f-4a0b-a1ba-fdee7eeed869"). InnerVolumeSpecName "kube-api-access-jnq9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.783027 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab331b3a-89a2-4d30-8c93-eec6b456d9be-kube-api-access-ns9m6" (OuterVolumeSpecName: "kube-api-access-ns9m6") pod "ab331b3a-89a2-4d30-8c93-eec6b456d9be" (UID: "ab331b3a-89a2-4d30-8c93-eec6b456d9be"). InnerVolumeSpecName "kube-api-access-ns9m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.879790 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns9m6\" (UniqueName: \"kubernetes.io/projected/ab331b3a-89a2-4d30-8c93-eec6b456d9be-kube-api-access-ns9m6\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.879846 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.879860 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/618c6158-f35f-4a0b-a1ba-fdee7eeed869-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.879889 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab331b3a-89a2-4d30-8c93-eec6b456d9be-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.879899 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnq9s\" (UniqueName: \"kubernetes.io/projected/618c6158-f35f-4a0b-a1ba-fdee7eeed869-kube-api-access-jnq9s\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.895859 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-f6rhs"] Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.898410 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b75798c-q5gqg" event={"ID":"618c6158-f35f-4a0b-a1ba-fdee7eeed869","Type":"ContainerDied","Data":"0935f0085565b181835a7d8472a49012eb1008b1b5307adfdda1dd5b561b533c"} Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.898502 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b75798c-q5gqg" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.905282 4966 generic.go:334] "Generic (PLEG): container finished" podID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" containerID="bedf75cee732948d22da311d019ae6801dbfae573135c678db59f60538a5f542" exitCode=0 Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.906265 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" event={"ID":"e3ae48f4-843d-4c46-9fd7-5f27f10863d1","Type":"ContainerDied","Data":"bedf75cee732948d22da311d019ae6801dbfae573135c678db59f60538a5f542"} Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.911459 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" event={"ID":"3e81f86f-d072-4057-807d-81d017f072f9","Type":"ContainerStarted","Data":"1ca1f40e30419dcae865d389041d39c16f770b817813aaaf6b8415cfc0e53fbf"} Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.911644 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.923753 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6d362072-2095-4d41-8ab5-7681c040faf6","Type":"ContainerStarted","Data":"9a0e34856fdff0066c6f34bd0e1e895a7987657471192c671fef7ec03f0acc12"} Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.948947 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b75798c-q5gqg"] Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.959407 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"83bca014-fd3a-430d-b532-77c1429a3728","Type":"ContainerStarted","Data":"55ce19ec12e9339a7619c2a2cd01b2b5be3bdead86217c2a66ce784ac7cfff53"} Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.959738 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b75798c-q5gqg"] Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.982116 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b562eab8-f0a5-4de7-a251-367555da04bf","Type":"ContainerStarted","Data":"334cbbc46e1527c1b055427cdc5fa442fa296bf76699481cd86d5e87d75e4d2f"} Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.986169 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" podStartSLOduration=9.019189737 podStartE2EDuration="22.986150065s" podCreationTimestamp="2025-12-17 08:37:08 +0000 UTC" firstStartedPulling="2025-12-17 08:37:15.332734512 +0000 UTC m=+970.877804464" lastFinishedPulling="2025-12-17 08:37:29.29969485 +0000 UTC m=+984.844764792" observedRunningTime="2025-12-17 08:37:30.950550869 +0000 UTC m=+986.495620811" watchObservedRunningTime="2025-12-17 08:37:30.986150065 +0000 UTC m=+986.531220017" Dec 17 08:37:30 crc kubenswrapper[4966]: I1217 08:37:30.988047 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bdca14dc-79c4-4e52-a748-9977656f7821","Type":"ContainerStarted","Data":"f4e0b3e58bf71c2d2ef5bd2388e9ada27432f48abe12780d08689b80a5997f1c"} Dec 17 08:37:31 crc kubenswrapper[4966]: I1217 08:37:30.999996 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" event={"ID":"ab331b3a-89a2-4d30-8c93-eec6b456d9be","Type":"ContainerDied","Data":"c0ccdab5a61fb118adbca23c1cace00a29969c42aa3972c3983e73e9da486dff"} Dec 17 08:37:31 crc kubenswrapper[4966]: I1217 08:37:31.000252 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64d8cd5c57-stcsp" Dec 17 08:37:31 crc kubenswrapper[4966]: I1217 08:37:31.004667 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cbln8" event={"ID":"3cb36f72-ce5e-4490-9009-7fe4443152b5","Type":"ContainerStarted","Data":"5d95e596ffe1dd62462017dc65266c898b889d8b9ad2037cac9dd9ce0773a43e"} Dec 17 08:37:31 crc kubenswrapper[4966]: I1217 08:37:31.007539 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb","Type":"ContainerStarted","Data":"4f8932c4782c2bda7d35a0975dd831c49dafc9e0b31ab55f6bd32af612a1a7bc"} Dec 17 08:37:31 crc kubenswrapper[4966]: I1217 08:37:31.058190 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64d8cd5c57-stcsp"] Dec 17 08:37:31 crc kubenswrapper[4966]: I1217 08:37:31.073839 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64d8cd5c57-stcsp"] Dec 17 08:37:31 crc kubenswrapper[4966]: W1217 08:37:31.169964 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30d8ef9a_31d4_4d75_91d2_cbf741bcf41f.slice/crio-4fe3f2e6cbb5283e8f616e23ec4a59b47a8a03cadbe10453413f35b500665835 WatchSource:0}: Error finding container 4fe3f2e6cbb5283e8f616e23ec4a59b47a8a03cadbe10453413f35b500665835: Status 404 returned error can't find the container with id 4fe3f2e6cbb5283e8f616e23ec4a59b47a8a03cadbe10453413f35b500665835 Dec 17 08:37:31 crc kubenswrapper[4966]: I1217 08:37:31.702936 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 17 08:37:31 crc kubenswrapper[4966]: W1217 08:37:31.899462 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0279d039_e449_4958_aebd_2d619653460d.slice/crio-58c11ee8688fd9e88d259a6f4322e5d255f35cb7311e92d0c0a3f4a84db56242 WatchSource:0}: Error finding container 58c11ee8688fd9e88d259a6f4322e5d255f35cb7311e92d0c0a3f4a84db56242: Status 404 returned error can't find the container with id 58c11ee8688fd9e88d259a6f4322e5d255f35cb7311e92d0c0a3f4a84db56242 Dec 17 08:37:31 crc kubenswrapper[4966]: I1217 08:37:31.916432 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 17 08:37:32 crc kubenswrapper[4966]: I1217 08:37:32.017287 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"0279d039-e449-4958-aebd-2d619653460d","Type":"ContainerStarted","Data":"58c11ee8688fd9e88d259a6f4322e5d255f35cb7311e92d0c0a3f4a84db56242"} Dec 17 08:37:32 crc kubenswrapper[4966]: I1217 08:37:32.020958 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f6rhs" event={"ID":"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f","Type":"ContainerStarted","Data":"4fe3f2e6cbb5283e8f616e23ec4a59b47a8a03cadbe10453413f35b500665835"} Dec 17 08:37:32 crc kubenswrapper[4966]: I1217 08:37:32.840519 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="618c6158-f35f-4a0b-a1ba-fdee7eeed869" path="/var/lib/kubelet/pods/618c6158-f35f-4a0b-a1ba-fdee7eeed869/volumes" Dec 17 08:37:32 crc kubenswrapper[4966]: I1217 08:37:32.840915 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab331b3a-89a2-4d30-8c93-eec6b456d9be" path="/var/lib/kubelet/pods/ab331b3a-89a2-4d30-8c93-eec6b456d9be/volumes" Dec 17 08:37:35 crc kubenswrapper[4966]: I1217 08:37:35.044800 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d7042a58-a47b-4eed-b4d6-625a3f57fe94","Type":"ContainerStarted","Data":"e460550863c4f1163678b6cd5e1664adc5d9e108ac3dd0da36776cdccebb8f5c"} Dec 17 08:37:38 crc kubenswrapper[4966]: I1217 08:37:38.757834 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:37:38 crc kubenswrapper[4966]: I1217 08:37:38.856705 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78d6cd766f-w5jdf"] Dec 17 08:37:39 crc kubenswrapper[4966]: I1217 08:37:39.080752 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" event={"ID":"e3ae48f4-843d-4c46-9fd7-5f27f10863d1","Type":"ContainerStarted","Data":"1d7ca923bdc345e4623b6be16e9d5316bebecad1c767b902b63099e0d1795adf"} Dec 17 08:37:39 crc kubenswrapper[4966]: I1217 08:37:39.080996 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" podUID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" containerName="dnsmasq-dns" containerID="cri-o://1d7ca923bdc345e4623b6be16e9d5316bebecad1c767b902b63099e0d1795adf" gracePeriod=10 Dec 17 08:37:39 crc kubenswrapper[4966]: I1217 08:37:39.081088 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.092068 4966 generic.go:334] "Generic (PLEG): container finished" podID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" containerID="1d7ca923bdc345e4623b6be16e9d5316bebecad1c767b902b63099e0d1795adf" exitCode=0 Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.092150 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" event={"ID":"e3ae48f4-843d-4c46-9fd7-5f27f10863d1","Type":"ContainerDied","Data":"1d7ca923bdc345e4623b6be16e9d5316bebecad1c767b902b63099e0d1795adf"} Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.528796 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.682003 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x99b9\" (UniqueName: \"kubernetes.io/projected/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-kube-api-access-x99b9\") pod \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.682454 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-config\") pod \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.682506 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-dns-svc\") pod \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\" (UID: \"e3ae48f4-843d-4c46-9fd7-5f27f10863d1\") " Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.686788 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-kube-api-access-x99b9" (OuterVolumeSpecName: "kube-api-access-x99b9") pod "e3ae48f4-843d-4c46-9fd7-5f27f10863d1" (UID: "e3ae48f4-843d-4c46-9fd7-5f27f10863d1"). InnerVolumeSpecName "kube-api-access-x99b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.725614 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e3ae48f4-843d-4c46-9fd7-5f27f10863d1" (UID: "e3ae48f4-843d-4c46-9fd7-5f27f10863d1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.726641 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-config" (OuterVolumeSpecName: "config") pod "e3ae48f4-843d-4c46-9fd7-5f27f10863d1" (UID: "e3ae48f4-843d-4c46-9fd7-5f27f10863d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.786151 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.786189 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:40 crc kubenswrapper[4966]: I1217 08:37:40.786201 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x99b9\" (UniqueName: \"kubernetes.io/projected/e3ae48f4-843d-4c46-9fd7-5f27f10863d1-kube-api-access-x99b9\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:41 crc kubenswrapper[4966]: I1217 08:37:41.101761 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" event={"ID":"e3ae48f4-843d-4c46-9fd7-5f27f10863d1","Type":"ContainerDied","Data":"4be211082464b8bb8c84e8d40b8ae6e5e4800d276ef10d4abb40d1c03ebf6b99"} Dec 17 08:37:41 crc kubenswrapper[4966]: I1217 08:37:41.101908 4966 scope.go:117] "RemoveContainer" containerID="1d7ca923bdc345e4623b6be16e9d5316bebecad1c767b902b63099e0d1795adf" Dec 17 08:37:41 crc kubenswrapper[4966]: I1217 08:37:41.101852 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78d6cd766f-w5jdf" Dec 17 08:37:41 crc kubenswrapper[4966]: I1217 08:37:41.128532 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78d6cd766f-w5jdf"] Dec 17 08:37:41 crc kubenswrapper[4966]: I1217 08:37:41.135020 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78d6cd766f-w5jdf"] Dec 17 08:37:41 crc kubenswrapper[4966]: I1217 08:37:41.194103 4966 scope.go:117] "RemoveContainer" containerID="bedf75cee732948d22da311d019ae6801dbfae573135c678db59f60538a5f542" Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.116775 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cbln8" event={"ID":"3cb36f72-ce5e-4490-9009-7fe4443152b5","Type":"ContainerStarted","Data":"9ad29a12d755679319617ca8a381124578ae694d8eeba25355e50a007eb2a92a"} Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.117400 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-cbln8" Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.121235 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6d362072-2095-4d41-8ab5-7681c040faf6","Type":"ContainerStarted","Data":"e44c7076367f8a48306618dfd21e589da8add83b19744abb2ed0c403591ae957"} Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.121277 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.123127 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"83bca014-fd3a-430d-b532-77c1429a3728","Type":"ContainerStarted","Data":"4fec3cc45062d820e28d7758bfc87a4e895ad2dbb6ec5c3f413853ca9320c6be"} Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.127267 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b562eab8-f0a5-4de7-a251-367555da04bf","Type":"ContainerStarted","Data":"779b84aab4577b94291a6ace7135ecec9a53f48d18f16e091015a95ab2a8e383"} Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.135911 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d7042a58-a47b-4eed-b4d6-625a3f57fe94","Type":"ContainerStarted","Data":"951fccecd9066cbc07404397e71a89d7f0846f7aac923df1e20eb3cf43e33795"} Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.141382 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f6rhs" event={"ID":"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f","Type":"ContainerStarted","Data":"aa4d4bd16212e1ba24328eacaee648f4a70e6cd777bfd0f1f0fee5094bd9617b"} Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.143065 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-cbln8" podStartSLOduration=14.254797168 podStartE2EDuration="24.143049692s" podCreationTimestamp="2025-12-17 08:37:18 +0000 UTC" firstStartedPulling="2025-12-17 08:37:30.55324947 +0000 UTC m=+986.098319412" lastFinishedPulling="2025-12-17 08:37:40.441501994 +0000 UTC m=+995.986571936" observedRunningTime="2025-12-17 08:37:42.140985526 +0000 UTC m=+997.686055468" watchObservedRunningTime="2025-12-17 08:37:42.143049692 +0000 UTC m=+997.688119634" Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.147226 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"0279d039-e449-4958-aebd-2d619653460d","Type":"ContainerStarted","Data":"57b6a352c776f755b3eabd262c1ce165dd16b6ebf72dff74ee7335152e5d1776"} Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.208683 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=21.107906072 podStartE2EDuration="30.208665471s" podCreationTimestamp="2025-12-17 08:37:12 +0000 UTC" firstStartedPulling="2025-12-17 08:37:30.330073722 +0000 UTC m=+985.875143664" lastFinishedPulling="2025-12-17 08:37:39.430833121 +0000 UTC m=+994.975903063" observedRunningTime="2025-12-17 08:37:42.205511275 +0000 UTC m=+997.750581217" watchObservedRunningTime="2025-12-17 08:37:42.208665471 +0000 UTC m=+997.753735413" Dec 17 08:37:42 crc kubenswrapper[4966]: I1217 08:37:42.839915 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" path="/var/lib/kubelet/pods/e3ae48f4-843d-4c46-9fd7-5f27f10863d1/volumes" Dec 17 08:37:43 crc kubenswrapper[4966]: I1217 08:37:43.154389 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1bd01548-3d59-4472-b8ef-a899b2cedee8","Type":"ContainerStarted","Data":"f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd"} Dec 17 08:37:43 crc kubenswrapper[4966]: I1217 08:37:43.157916 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb","Type":"ContainerStarted","Data":"40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a"} Dec 17 08:37:43 crc kubenswrapper[4966]: I1217 08:37:43.160086 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bdca14dc-79c4-4e52-a748-9977656f7821","Type":"ContainerStarted","Data":"36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13"} Dec 17 08:37:43 crc kubenswrapper[4966]: I1217 08:37:43.160925 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 17 08:37:43 crc kubenswrapper[4966]: I1217 08:37:43.162215 4966 generic.go:334] "Generic (PLEG): container finished" podID="30d8ef9a-31d4-4d75-91d2-cbf741bcf41f" containerID="aa4d4bd16212e1ba24328eacaee648f4a70e6cd777bfd0f1f0fee5094bd9617b" exitCode=0 Dec 17 08:37:43 crc kubenswrapper[4966]: I1217 08:37:43.163930 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f6rhs" event={"ID":"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f","Type":"ContainerDied","Data":"aa4d4bd16212e1ba24328eacaee648f4a70e6cd777bfd0f1f0fee5094bd9617b"} Dec 17 08:37:43 crc kubenswrapper[4966]: I1217 08:37:43.268315 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=18.172618465 podStartE2EDuration="29.268299715s" podCreationTimestamp="2025-12-17 08:37:14 +0000 UTC" firstStartedPulling="2025-12-17 08:37:30.29606556 +0000 UTC m=+985.841135502" lastFinishedPulling="2025-12-17 08:37:41.39174681 +0000 UTC m=+996.936816752" observedRunningTime="2025-12-17 08:37:43.261496609 +0000 UTC m=+998.806566551" watchObservedRunningTime="2025-12-17 08:37:43.268299715 +0000 UTC m=+998.813369657" Dec 17 08:37:44 crc kubenswrapper[4966]: I1217 08:37:44.177812 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f6rhs" event={"ID":"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f","Type":"ContainerStarted","Data":"1f6267fd7122cf804ea58b5d9452bab455ed4075353676b2e61d82fd644d3cf3"} Dec 17 08:37:44 crc kubenswrapper[4966]: I1217 08:37:44.178102 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f6rhs" event={"ID":"30d8ef9a-31d4-4d75-91d2-cbf741bcf41f","Type":"ContainerStarted","Data":"a16800dd2cc6f16d7a2459d6c443186801f72371d093a7b4e00b472daca71a45"} Dec 17 08:37:44 crc kubenswrapper[4966]: I1217 08:37:44.178795 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:44 crc kubenswrapper[4966]: I1217 08:37:44.178822 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:37:44 crc kubenswrapper[4966]: I1217 08:37:44.201721 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-f6rhs" podStartSLOduration=16.936467552 podStartE2EDuration="26.201703789s" podCreationTimestamp="2025-12-17 08:37:18 +0000 UTC" firstStartedPulling="2025-12-17 08:37:31.175222688 +0000 UTC m=+986.720292630" lastFinishedPulling="2025-12-17 08:37:40.440458935 +0000 UTC m=+995.985528867" observedRunningTime="2025-12-17 08:37:44.197736981 +0000 UTC m=+999.742806923" watchObservedRunningTime="2025-12-17 08:37:44.201703789 +0000 UTC m=+999.746773731" Dec 17 08:37:46 crc kubenswrapper[4966]: I1217 08:37:46.807482 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:37:46 crc kubenswrapper[4966]: I1217 08:37:46.807996 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:37:47 crc kubenswrapper[4966]: I1217 08:37:47.199153 4966 generic.go:334] "Generic (PLEG): container finished" podID="b562eab8-f0a5-4de7-a251-367555da04bf" containerID="779b84aab4577b94291a6ace7135ecec9a53f48d18f16e091015a95ab2a8e383" exitCode=0 Dec 17 08:37:47 crc kubenswrapper[4966]: I1217 08:37:47.199252 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b562eab8-f0a5-4de7-a251-367555da04bf","Type":"ContainerDied","Data":"779b84aab4577b94291a6ace7135ecec9a53f48d18f16e091015a95ab2a8e383"} Dec 17 08:37:47 crc kubenswrapper[4966]: I1217 08:37:47.202244 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d7042a58-a47b-4eed-b4d6-625a3f57fe94","Type":"ContainerStarted","Data":"e5668c83efff57eb3f9e7f0dfc35eb73030faa5f9010372c764d61f9ea2d3b0f"} Dec 17 08:37:47 crc kubenswrapper[4966]: I1217 08:37:47.206379 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"0279d039-e449-4958-aebd-2d619653460d","Type":"ContainerStarted","Data":"1e2854916ffcbd7f1b5b01dc26d9b0b9802df6a3460ee98888df34e5af114689"} Dec 17 08:37:47 crc kubenswrapper[4966]: I1217 08:37:47.210440 4966 generic.go:334] "Generic (PLEG): container finished" podID="83bca014-fd3a-430d-b532-77c1429a3728" containerID="4fec3cc45062d820e28d7758bfc87a4e895ad2dbb6ec5c3f413853ca9320c6be" exitCode=0 Dec 17 08:37:47 crc kubenswrapper[4966]: I1217 08:37:47.210475 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"83bca014-fd3a-430d-b532-77c1429a3728","Type":"ContainerDied","Data":"4fec3cc45062d820e28d7758bfc87a4e895ad2dbb6ec5c3f413853ca9320c6be"} Dec 17 08:37:47 crc kubenswrapper[4966]: I1217 08:37:47.268079 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=14.417196408 podStartE2EDuration="26.268057177s" podCreationTimestamp="2025-12-17 08:37:21 +0000 UTC" firstStartedPulling="2025-12-17 08:37:34.583308932 +0000 UTC m=+990.128378894" lastFinishedPulling="2025-12-17 08:37:46.434169711 +0000 UTC m=+1001.979239663" observedRunningTime="2025-12-17 08:37:47.266221407 +0000 UTC m=+1002.811291379" watchObservedRunningTime="2025-12-17 08:37:47.268057177 +0000 UTC m=+1002.813127139" Dec 17 08:37:47 crc kubenswrapper[4966]: I1217 08:37:47.328540 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.808554477 podStartE2EDuration="29.328518185s" podCreationTimestamp="2025-12-17 08:37:18 +0000 UTC" firstStartedPulling="2025-12-17 08:37:31.902614865 +0000 UTC m=+987.447684807" lastFinishedPulling="2025-12-17 08:37:46.422578573 +0000 UTC m=+1001.967648515" observedRunningTime="2025-12-17 08:37:47.324239878 +0000 UTC m=+1002.869309820" watchObservedRunningTime="2025-12-17 08:37:47.328518185 +0000 UTC m=+1002.873588127" Dec 17 08:37:47 crc kubenswrapper[4966]: I1217 08:37:47.592154 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:48 crc kubenswrapper[4966]: I1217 08:37:48.125991 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 17 08:37:48 crc kubenswrapper[4966]: I1217 08:37:48.255817 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"83bca014-fd3a-430d-b532-77c1429a3728","Type":"ContainerStarted","Data":"08d3ca697c5979812f11ab0b67b757af9a10e4918410d3a558cc1e1882fdd25a"} Dec 17 08:37:48 crc kubenswrapper[4966]: I1217 08:37:48.264930 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b562eab8-f0a5-4de7-a251-367555da04bf","Type":"ContainerStarted","Data":"57b729781d546c49c7f5b0339afd1337b4afb41d4062f2680f976ce599322d81"} Dec 17 08:37:48 crc kubenswrapper[4966]: I1217 08:37:48.302184 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=27.759039538 podStartE2EDuration="37.302158772s" podCreationTimestamp="2025-12-17 08:37:11 +0000 UTC" firstStartedPulling="2025-12-17 08:37:30.330296598 +0000 UTC m=+985.875366540" lastFinishedPulling="2025-12-17 08:37:39.873415832 +0000 UTC m=+995.418485774" observedRunningTime="2025-12-17 08:37:48.280613842 +0000 UTC m=+1003.825683784" watchObservedRunningTime="2025-12-17 08:37:48.302158772 +0000 UTC m=+1003.847228714" Dec 17 08:37:48 crc kubenswrapper[4966]: I1217 08:37:48.320068 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=28.209985159 podStartE2EDuration="38.320041052s" podCreationTimestamp="2025-12-17 08:37:10 +0000 UTC" firstStartedPulling="2025-12-17 08:37:30.330559886 +0000 UTC m=+985.875629828" lastFinishedPulling="2025-12-17 08:37:40.440615779 +0000 UTC m=+995.985685721" observedRunningTime="2025-12-17 08:37:48.315392634 +0000 UTC m=+1003.860462576" watchObservedRunningTime="2025-12-17 08:37:48.320041052 +0000 UTC m=+1003.865110994" Dec 17 08:37:49 crc kubenswrapper[4966]: I1217 08:37:49.592314 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:49 crc kubenswrapper[4966]: I1217 08:37:49.601228 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:49 crc kubenswrapper[4966]: I1217 08:37:49.601275 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:49 crc kubenswrapper[4966]: I1217 08:37:49.655995 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:49 crc kubenswrapper[4966]: I1217 08:37:49.656092 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.312999 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.315785 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.480249 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79444cc97c-mqjzd"] Dec 17 08:37:50 crc kubenswrapper[4966]: E1217 08:37:50.480596 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" containerName="init" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.480612 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" containerName="init" Dec 17 08:37:50 crc kubenswrapper[4966]: E1217 08:37:50.480639 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" containerName="dnsmasq-dns" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.480645 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" containerName="dnsmasq-dns" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.480855 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ae48f4-843d-4c46-9fd7-5f27f10863d1" containerName="dnsmasq-dns" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.483284 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.486014 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.566021 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-ovsdbserver-sb\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.566093 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79444cc97c-mqjzd"] Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.566100 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-dns-svc\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.566238 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9d9c\" (UniqueName: \"kubernetes.io/projected/10b668b7-83f1-4fa9-8f77-0536a765d7ee-kube-api-access-g9d9c\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.566322 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-config\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.668530 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-ovsdbserver-sb\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.668587 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-dns-svc\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.668830 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9d9c\" (UniqueName: \"kubernetes.io/projected/10b668b7-83f1-4fa9-8f77-0536a765d7ee-kube-api-access-g9d9c\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.668937 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-config\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.669642 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-ovsdbserver-sb\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.669742 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-config\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.670201 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-dns-svc\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.687963 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9d9c\" (UniqueName: \"kubernetes.io/projected/10b668b7-83f1-4fa9-8f77-0536a765d7ee-kube-api-access-g9d9c\") pod \"dnsmasq-dns-79444cc97c-mqjzd\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:50 crc kubenswrapper[4966]: I1217 08:37:50.803151 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.382089 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79444cc97c-mqjzd"] Dec 17 08:37:51 crc kubenswrapper[4966]: W1217 08:37:51.399859 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10b668b7_83f1_4fa9_8f77_0536a765d7ee.slice/crio-60087d55b779cdbf8f8c9014c9acbbadebeb71bf714a2875462d685a314ada79 WatchSource:0}: Error finding container 60087d55b779cdbf8f8c9014c9acbbadebeb71bf714a2875462d685a314ada79: Status 404 returned error can't find the container with id 60087d55b779cdbf8f8c9014c9acbbadebeb71bf714a2875462d685a314ada79 Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.416936 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-fdx4l"] Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.418282 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.426376 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.428902 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.442898 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.447173 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.447413 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.447515 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.447627 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-zwsw9" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.459685 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fdx4l"] Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.504507 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-scripts\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.504856 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.504954 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.504979 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g92j\" (UniqueName: \"kubernetes.io/projected/356a494a-8dfe-4478-8486-9fc18d0eb8d4-kube-api-access-6g92j\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.505042 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/356a494a-8dfe-4478-8486-9fc18d0eb8d4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.505085 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/356a494a-8dfe-4478-8486-9fc18d0eb8d4-config\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.505116 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356a494a-8dfe-4478-8486-9fc18d0eb8d4-combined-ca-bundle\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.505188 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.505311 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgbjl\" (UniqueName: \"kubernetes.io/projected/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-kube-api-access-fgbjl\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.505516 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/356a494a-8dfe-4478-8486-9fc18d0eb8d4-ovs-rundir\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.505725 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/356a494a-8dfe-4478-8486-9fc18d0eb8d4-ovn-rundir\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.506604 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.506843 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-config\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.538068 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.543941 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79444cc97c-mqjzd"] Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.590943 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56c5cf55c-46stc"] Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.592632 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.594294 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.594487 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56c5cf55c-46stc"] Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610280 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/356a494a-8dfe-4478-8486-9fc18d0eb8d4-ovs-rundir\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610333 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/356a494a-8dfe-4478-8486-9fc18d0eb8d4-ovn-rundir\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610351 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610389 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-config\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610425 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-scripts\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610441 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610468 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610485 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g92j\" (UniqueName: \"kubernetes.io/projected/356a494a-8dfe-4478-8486-9fc18d0eb8d4-kube-api-access-6g92j\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610506 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/356a494a-8dfe-4478-8486-9fc18d0eb8d4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610526 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/356a494a-8dfe-4478-8486-9fc18d0eb8d4-config\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610542 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356a494a-8dfe-4478-8486-9fc18d0eb8d4-combined-ca-bundle\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610579 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.610619 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgbjl\" (UniqueName: \"kubernetes.io/projected/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-kube-api-access-fgbjl\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.611171 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/356a494a-8dfe-4478-8486-9fc18d0eb8d4-ovs-rundir\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.611222 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/356a494a-8dfe-4478-8486-9fc18d0eb8d4-ovn-rundir\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.613299 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.613439 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-config\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.614122 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-scripts\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.614465 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/356a494a-8dfe-4478-8486-9fc18d0eb8d4-config\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.625952 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.628123 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.630417 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.632472 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/356a494a-8dfe-4478-8486-9fc18d0eb8d4-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.637956 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.637998 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.638686 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/356a494a-8dfe-4478-8486-9fc18d0eb8d4-combined-ca-bundle\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.641060 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgbjl\" (UniqueName: \"kubernetes.io/projected/ece61b7a-78e0-4e28-9f6d-62c89db6c01d-kube-api-access-fgbjl\") pod \"ovn-northd-0\" (UID: \"ece61b7a-78e0-4e28-9f6d-62c89db6c01d\") " pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.643273 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g92j\" (UniqueName: \"kubernetes.io/projected/356a494a-8dfe-4478-8486-9fc18d0eb8d4-kube-api-access-6g92j\") pod \"ovn-controller-metrics-fdx4l\" (UID: \"356a494a-8dfe-4478-8486-9fc18d0eb8d4\") " pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.712059 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clg44\" (UniqueName: \"kubernetes.io/projected/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-kube-api-access-clg44\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.713400 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-nb\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.713633 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-config\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.713773 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-sb\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.714048 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-dns-svc\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.815299 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-fdx4l" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.815723 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-nb\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.815789 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-config\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.815834 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-sb\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.815922 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-dns-svc\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.816012 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clg44\" (UniqueName: \"kubernetes.io/projected/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-kube-api-access-clg44\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.817829 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-dns-svc\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.817893 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-sb\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.817893 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-nb\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.831102 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.831561 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-config\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.847011 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clg44\" (UniqueName: \"kubernetes.io/projected/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-kube-api-access-clg44\") pod \"dnsmasq-dns-56c5cf55c-46stc\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:51 crc kubenswrapper[4966]: I1217 08:37:51.939600 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:52 crc kubenswrapper[4966]: I1217 08:37:52.323859 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" event={"ID":"10b668b7-83f1-4fa9-8f77-0536a765d7ee","Type":"ContainerStarted","Data":"df333390aa085c6233b99b45848dd729ff48e08d7abc166ec471081af572db50"} Dec 17 08:37:52 crc kubenswrapper[4966]: I1217 08:37:52.323935 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" event={"ID":"10b668b7-83f1-4fa9-8f77-0536a765d7ee","Type":"ContainerStarted","Data":"60087d55b779cdbf8f8c9014c9acbbadebeb71bf714a2875462d685a314ada79"} Dec 17 08:37:52 crc kubenswrapper[4966]: I1217 08:37:52.358551 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-fdx4l"] Dec 17 08:37:52 crc kubenswrapper[4966]: W1217 08:37:52.503977 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe914b29_6ca2_4c4d_a97b_6b5e1bc5ab22.slice/crio-a51bb661b058500dbc3b2650838ec0f1112b4c89e3d530ee50219e975e6e108f WatchSource:0}: Error finding container a51bb661b058500dbc3b2650838ec0f1112b4c89e3d530ee50219e975e6e108f: Status 404 returned error can't find the container with id a51bb661b058500dbc3b2650838ec0f1112b4c89e3d530ee50219e975e6e108f Dec 17 08:37:52 crc kubenswrapper[4966]: I1217 08:37:52.510778 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56c5cf55c-46stc"] Dec 17 08:37:52 crc kubenswrapper[4966]: W1217 08:37:52.512000 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podece61b7a_78e0_4e28_9f6d_62c89db6c01d.slice/crio-38d632cd6ebd176aa5bf142d84672272d8780e06198644ff100762289a46134c WatchSource:0}: Error finding container 38d632cd6ebd176aa5bf142d84672272d8780e06198644ff100762289a46134c: Status 404 returned error can't find the container with id 38d632cd6ebd176aa5bf142d84672272d8780e06198644ff100762289a46134c Dec 17 08:37:52 crc kubenswrapper[4966]: I1217 08:37:52.518022 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 17 08:37:52 crc kubenswrapper[4966]: I1217 08:37:52.783313 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:52 crc kubenswrapper[4966]: I1217 08:37:52.783698 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.334781 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fdx4l" event={"ID":"356a494a-8dfe-4478-8486-9fc18d0eb8d4","Type":"ContainerStarted","Data":"8ce06795de0c0e64d3c9e1f996aad157f5c00f7eae7c9ef8d52cea6d127518c7"} Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.335277 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-fdx4l" event={"ID":"356a494a-8dfe-4478-8486-9fc18d0eb8d4","Type":"ContainerStarted","Data":"3eb0f44aeed6206f4c9dcd474200fe8901ed915ec75bdee83881314a1777a563"} Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.336539 4966 generic.go:334] "Generic (PLEG): container finished" podID="be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" containerID="add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45" exitCode=0 Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.336621 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" event={"ID":"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22","Type":"ContainerDied","Data":"add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45"} Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.336639 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" event={"ID":"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22","Type":"ContainerStarted","Data":"a51bb661b058500dbc3b2650838ec0f1112b4c89e3d530ee50219e975e6e108f"} Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.338078 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ece61b7a-78e0-4e28-9f6d-62c89db6c01d","Type":"ContainerStarted","Data":"38d632cd6ebd176aa5bf142d84672272d8780e06198644ff100762289a46134c"} Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.339830 4966 generic.go:334] "Generic (PLEG): container finished" podID="10b668b7-83f1-4fa9-8f77-0536a765d7ee" containerID="df333390aa085c6233b99b45848dd729ff48e08d7abc166ec471081af572db50" exitCode=0 Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.339947 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" event={"ID":"10b668b7-83f1-4fa9-8f77-0536a765d7ee","Type":"ContainerDied","Data":"df333390aa085c6233b99b45848dd729ff48e08d7abc166ec471081af572db50"} Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.356280 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-fdx4l" podStartSLOduration=2.3562573430000002 podStartE2EDuration="2.356257343s" podCreationTimestamp="2025-12-17 08:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:37:53.352900741 +0000 UTC m=+1008.897970683" watchObservedRunningTime="2025-12-17 08:37:53.356257343 +0000 UTC m=+1008.901327285" Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.879389 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.970510 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-dns-svc\") pod \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.970575 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-config\") pod \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.970619 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9d9c\" (UniqueName: \"kubernetes.io/projected/10b668b7-83f1-4fa9-8f77-0536a765d7ee-kube-api-access-g9d9c\") pod \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.970660 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-ovsdbserver-sb\") pod \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\" (UID: \"10b668b7-83f1-4fa9-8f77-0536a765d7ee\") " Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.976537 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10b668b7-83f1-4fa9-8f77-0536a765d7ee-kube-api-access-g9d9c" (OuterVolumeSpecName: "kube-api-access-g9d9c") pod "10b668b7-83f1-4fa9-8f77-0536a765d7ee" (UID: "10b668b7-83f1-4fa9-8f77-0536a765d7ee"). InnerVolumeSpecName "kube-api-access-g9d9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.989970 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-config" (OuterVolumeSpecName: "config") pod "10b668b7-83f1-4fa9-8f77-0536a765d7ee" (UID: "10b668b7-83f1-4fa9-8f77-0536a765d7ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:53 crc kubenswrapper[4966]: I1217 08:37:53.999824 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "10b668b7-83f1-4fa9-8f77-0536a765d7ee" (UID: "10b668b7-83f1-4fa9-8f77-0536a765d7ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.002322 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "10b668b7-83f1-4fa9-8f77-0536a765d7ee" (UID: "10b668b7-83f1-4fa9-8f77-0536a765d7ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.072683 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.072719 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.072731 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9d9c\" (UniqueName: \"kubernetes.io/projected/10b668b7-83f1-4fa9-8f77-0536a765d7ee-kube-api-access-g9d9c\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.072741 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10b668b7-83f1-4fa9-8f77-0536a765d7ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.350708 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ece61b7a-78e0-4e28-9f6d-62c89db6c01d","Type":"ContainerStarted","Data":"ae80a3246c0e7285fe307ead477fba72dcde2a85e17975ca012e88830754bcae"} Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.350765 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ece61b7a-78e0-4e28-9f6d-62c89db6c01d","Type":"ContainerStarted","Data":"2d889d4b93b0435325beaf62fd25559a65b472ab62095488554f9371e6df4703"} Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.350807 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.352778 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" event={"ID":"10b668b7-83f1-4fa9-8f77-0536a765d7ee","Type":"ContainerDied","Data":"60087d55b779cdbf8f8c9014c9acbbadebeb71bf714a2875462d685a314ada79"} Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.352808 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79444cc97c-mqjzd" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.352833 4966 scope.go:117] "RemoveContainer" containerID="df333390aa085c6233b99b45848dd729ff48e08d7abc166ec471081af572db50" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.356430 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" event={"ID":"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22","Type":"ContainerStarted","Data":"cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561"} Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.376865 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.133838777 podStartE2EDuration="3.376824337s" podCreationTimestamp="2025-12-17 08:37:51 +0000 UTC" firstStartedPulling="2025-12-17 08:37:52.514327276 +0000 UTC m=+1008.059397218" lastFinishedPulling="2025-12-17 08:37:53.757312836 +0000 UTC m=+1009.302382778" observedRunningTime="2025-12-17 08:37:54.372623802 +0000 UTC m=+1009.917693754" watchObservedRunningTime="2025-12-17 08:37:54.376824337 +0000 UTC m=+1009.921894289" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.441143 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79444cc97c-mqjzd"] Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.446644 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79444cc97c-mqjzd"] Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.456200 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" podStartSLOduration=3.456188181 podStartE2EDuration="3.456188181s" podCreationTimestamp="2025-12-17 08:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:37:54.453502748 +0000 UTC m=+1009.998572690" watchObservedRunningTime="2025-12-17 08:37:54.456188181 +0000 UTC m=+1010.001258113" Dec 17 08:37:54 crc kubenswrapper[4966]: I1217 08:37:54.852625 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10b668b7-83f1-4fa9-8f77-0536a765d7ee" path="/var/lib/kubelet/pods/10b668b7-83f1-4fa9-8f77-0536a765d7ee/volumes" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.318272 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.381102 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.398036 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.569108 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56c5cf55c-46stc"] Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.595304 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cbbc897ff-7tpcm"] Dec 17 08:37:55 crc kubenswrapper[4966]: E1217 08:37:55.595671 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10b668b7-83f1-4fa9-8f77-0536a765d7ee" containerName="init" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.595693 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="10b668b7-83f1-4fa9-8f77-0536a765d7ee" containerName="init" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.601904 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="10b668b7-83f1-4fa9-8f77-0536a765d7ee" containerName="init" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.603089 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.610843 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cbbc897ff-7tpcm"] Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.658292 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.699430 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-sb\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.699497 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-nb\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.699576 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-dns-svc\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.699613 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsdc7\" (UniqueName: \"kubernetes.io/projected/0c2b7867-860e-46c7-9981-55fefdfe7458-kube-api-access-xsdc7\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.699677 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-config\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.800661 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-dns-svc\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.800712 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsdc7\" (UniqueName: \"kubernetes.io/projected/0c2b7867-860e-46c7-9981-55fefdfe7458-kube-api-access-xsdc7\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.800786 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-config\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.800805 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-sb\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.800833 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-nb\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.801622 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-dns-svc\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.801655 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-nb\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.801677 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-sb\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.802232 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-config\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.825283 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsdc7\" (UniqueName: \"kubernetes.io/projected/0c2b7867-860e-46c7-9981-55fefdfe7458-kube-api-access-xsdc7\") pod \"dnsmasq-dns-5cbbc897ff-7tpcm\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:55 crc kubenswrapper[4966]: I1217 08:37:55.940959 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.399232 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cbbc897ff-7tpcm"] Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.784787 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.794622 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.804130 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.804564 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.804697 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-fds24" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.804837 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.817664 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.917068 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.917208 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnblp\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-kube-api-access-qnblp\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.917325 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.917351 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cbc21129-413d-4f88-9287-85c9a0fcfa68-lock\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:56 crc kubenswrapper[4966]: I1217 08:37:56.917585 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cbc21129-413d-4f88-9287-85c9a0fcfa68-cache\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.019230 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cbc21129-413d-4f88-9287-85c9a0fcfa68-cache\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.020036 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.020089 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnblp\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-kube-api-access-qnblp\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.020132 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.020623 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cbc21129-413d-4f88-9287-85c9a0fcfa68-lock\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.020498 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.019983 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cbc21129-413d-4f88-9287-85c9a0fcfa68-cache\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.021535 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cbc21129-413d-4f88-9287-85c9a0fcfa68-lock\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: E1217 08:37:57.020580 4966 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 17 08:37:57 crc kubenswrapper[4966]: E1217 08:37:57.021577 4966 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 17 08:37:57 crc kubenswrapper[4966]: E1217 08:37:57.021619 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift podName:cbc21129-413d-4f88-9287-85c9a0fcfa68 nodeName:}" failed. No retries permitted until 2025-12-17 08:37:57.52160283 +0000 UTC m=+1013.066672772 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift") pod "swift-storage-0" (UID: "cbc21129-413d-4f88-9287-85c9a0fcfa68") : configmap "swift-ring-files" not found Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.038897 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnblp\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-kube-api-access-qnblp\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.043928 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.393367 4966 generic.go:334] "Generic (PLEG): container finished" podID="0c2b7867-860e-46c7-9981-55fefdfe7458" containerID="e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1" exitCode=0 Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.393464 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" event={"ID":"0c2b7867-860e-46c7-9981-55fefdfe7458","Type":"ContainerDied","Data":"e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1"} Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.393519 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" event={"ID":"0c2b7867-860e-46c7-9981-55fefdfe7458","Type":"ContainerStarted","Data":"508bcc1a6fd61478a53ba81c31a7bc05d3126f5e103c9a42ca61d4077da7d02f"} Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.393544 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" podUID="be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" containerName="dnsmasq-dns" containerID="cri-o://cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561" gracePeriod=10 Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.528319 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:57 crc kubenswrapper[4966]: E1217 08:37:57.529184 4966 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 17 08:37:57 crc kubenswrapper[4966]: E1217 08:37:57.529207 4966 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 17 08:37:57 crc kubenswrapper[4966]: E1217 08:37:57.529272 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift podName:cbc21129-413d-4f88-9287-85c9a0fcfa68 nodeName:}" failed. No retries permitted until 2025-12-17 08:37:58.529258133 +0000 UTC m=+1014.074328075 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift") pod "swift-storage-0" (UID: "cbc21129-413d-4f88-9287-85c9a0fcfa68") : configmap "swift-ring-files" not found Dec 17 08:37:57 crc kubenswrapper[4966]: I1217 08:37:57.869585 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.036018 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-sb\") pod \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.036168 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clg44\" (UniqueName: \"kubernetes.io/projected/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-kube-api-access-clg44\") pod \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.036237 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-nb\") pod \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.036339 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-dns-svc\") pod \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.036761 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-config\") pod \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\" (UID: \"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22\") " Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.051103 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-kube-api-access-clg44" (OuterVolumeSpecName: "kube-api-access-clg44") pod "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" (UID: "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22"). InnerVolumeSpecName "kube-api-access-clg44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.078802 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" (UID: "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.097310 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" (UID: "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.098631 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-config" (OuterVolumeSpecName: "config") pod "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" (UID: "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.103345 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" (UID: "be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.145982 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.146032 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clg44\" (UniqueName: \"kubernetes.io/projected/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-kube-api-access-clg44\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.146043 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.146051 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.146064 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.391087 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-0937-account-create-update-bq6tm"] Dec 17 08:37:58 crc kubenswrapper[4966]: E1217 08:37:58.391407 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" containerName="init" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.391418 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" containerName="init" Dec 17 08:37:58 crc kubenswrapper[4966]: E1217 08:37:58.391446 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" containerName="dnsmasq-dns" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.391452 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" containerName="dnsmasq-dns" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.391605 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" containerName="dnsmasq-dns" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.392124 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.394624 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.401753 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0937-account-create-update-bq6tm"] Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.406527 4966 generic.go:334] "Generic (PLEG): container finished" podID="be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" containerID="cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561" exitCode=0 Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.406578 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" event={"ID":"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22","Type":"ContainerDied","Data":"cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561"} Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.406600 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" event={"ID":"be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22","Type":"ContainerDied","Data":"a51bb661b058500dbc3b2650838ec0f1112b4c89e3d530ee50219e975e6e108f"} Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.406618 4966 scope.go:117] "RemoveContainer" containerID="cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.406761 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56c5cf55c-46stc" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.410968 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" event={"ID":"0c2b7867-860e-46c7-9981-55fefdfe7458","Type":"ContainerStarted","Data":"2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d"} Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.429772 4966 scope.go:117] "RemoveContainer" containerID="add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.438889 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-6l529"] Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.439944 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6l529" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.454468 4966 scope.go:117] "RemoveContainer" containerID="cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561" Dec 17 08:37:58 crc kubenswrapper[4966]: E1217 08:37:58.456742 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561\": container with ID starting with cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561 not found: ID does not exist" containerID="cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.456783 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561"} err="failed to get container status \"cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561\": rpc error: code = NotFound desc = could not find container \"cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561\": container with ID starting with cd9e26f7fd59ca92be9231c931d98e100259202f394807249c4dfc26507b4561 not found: ID does not exist" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.456805 4966 scope.go:117] "RemoveContainer" containerID="add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45" Dec 17 08:37:58 crc kubenswrapper[4966]: E1217 08:37:58.459845 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45\": container with ID starting with add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45 not found: ID does not exist" containerID="add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.459902 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45"} err="failed to get container status \"add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45\": rpc error: code = NotFound desc = could not find container \"add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45\": container with ID starting with add50d615e85e5fa9691345590d0ca97d3735feffb16a27089f8aa0f6117dd45 not found: ID does not exist" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.469237 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6l529"] Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.470514 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" podStartSLOduration=3.470502552 podStartE2EDuration="3.470502552s" podCreationTimestamp="2025-12-17 08:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:37:58.4689347 +0000 UTC m=+1014.014004652" watchObservedRunningTime="2025-12-17 08:37:58.470502552 +0000 UTC m=+1014.015572494" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.495935 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56c5cf55c-46stc"] Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.503283 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56c5cf55c-46stc"] Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.551014 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csbm9\" (UniqueName: \"kubernetes.io/projected/3153dd6b-a60e-4da6-8463-a6e7b86c0277-kube-api-access-csbm9\") pod \"glance-db-create-6l529\" (UID: \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\") " pod="openstack/glance-db-create-6l529" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.551061 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d6c04f-29ad-41b5-a6b7-a8868cf02155-operator-scripts\") pod \"glance-0937-account-create-update-bq6tm\" (UID: \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\") " pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.551082 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xxqj\" (UniqueName: \"kubernetes.io/projected/84d6c04f-29ad-41b5-a6b7-a8868cf02155-kube-api-access-8xxqj\") pod \"glance-0937-account-create-update-bq6tm\" (UID: \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\") " pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.551423 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.551518 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3153dd6b-a60e-4da6-8463-a6e7b86c0277-operator-scripts\") pod \"glance-db-create-6l529\" (UID: \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\") " pod="openstack/glance-db-create-6l529" Dec 17 08:37:58 crc kubenswrapper[4966]: E1217 08:37:58.552079 4966 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 17 08:37:58 crc kubenswrapper[4966]: E1217 08:37:58.552105 4966 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 17 08:37:58 crc kubenswrapper[4966]: E1217 08:37:58.552157 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift podName:cbc21129-413d-4f88-9287-85c9a0fcfa68 nodeName:}" failed. No retries permitted until 2025-12-17 08:38:00.552135531 +0000 UTC m=+1016.097205473 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift") pod "swift-storage-0" (UID: "cbc21129-413d-4f88-9287-85c9a0fcfa68") : configmap "swift-ring-files" not found Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.652791 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csbm9\" (UniqueName: \"kubernetes.io/projected/3153dd6b-a60e-4da6-8463-a6e7b86c0277-kube-api-access-csbm9\") pod \"glance-db-create-6l529\" (UID: \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\") " pod="openstack/glance-db-create-6l529" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.652842 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d6c04f-29ad-41b5-a6b7-a8868cf02155-operator-scripts\") pod \"glance-0937-account-create-update-bq6tm\" (UID: \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\") " pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.652863 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xxqj\" (UniqueName: \"kubernetes.io/projected/84d6c04f-29ad-41b5-a6b7-a8868cf02155-kube-api-access-8xxqj\") pod \"glance-0937-account-create-update-bq6tm\" (UID: \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\") " pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.653031 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3153dd6b-a60e-4da6-8463-a6e7b86c0277-operator-scripts\") pod \"glance-db-create-6l529\" (UID: \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\") " pod="openstack/glance-db-create-6l529" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.653764 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3153dd6b-a60e-4da6-8463-a6e7b86c0277-operator-scripts\") pod \"glance-db-create-6l529\" (UID: \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\") " pod="openstack/glance-db-create-6l529" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.653828 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d6c04f-29ad-41b5-a6b7-a8868cf02155-operator-scripts\") pod \"glance-0937-account-create-update-bq6tm\" (UID: \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\") " pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.671264 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xxqj\" (UniqueName: \"kubernetes.io/projected/84d6c04f-29ad-41b5-a6b7-a8868cf02155-kube-api-access-8xxqj\") pod \"glance-0937-account-create-update-bq6tm\" (UID: \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\") " pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.671325 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csbm9\" (UniqueName: \"kubernetes.io/projected/3153dd6b-a60e-4da6-8463-a6e7b86c0277-kube-api-access-csbm9\") pod \"glance-db-create-6l529\" (UID: \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\") " pod="openstack/glance-db-create-6l529" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.705734 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.795174 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6l529" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.851468 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22" path="/var/lib/kubelet/pods/be914b29-6ca2-4c4d-a97b-6b5e1bc5ab22/volumes" Dec 17 08:37:58 crc kubenswrapper[4966]: I1217 08:37:58.937191 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:59 crc kubenswrapper[4966]: I1217 08:37:59.327459 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 17 08:37:59 crc kubenswrapper[4966]: I1217 08:37:59.336292 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0937-account-create-update-bq6tm"] Dec 17 08:37:59 crc kubenswrapper[4966]: I1217 08:37:59.348560 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6l529"] Dec 17 08:37:59 crc kubenswrapper[4966]: I1217 08:37:59.440751 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0937-account-create-update-bq6tm" event={"ID":"84d6c04f-29ad-41b5-a6b7-a8868cf02155","Type":"ContainerStarted","Data":"aa22da15c1e0ac91e15497394f386426ea54867a58a3370e0517af07085aa49c"} Dec 17 08:37:59 crc kubenswrapper[4966]: I1217 08:37:59.459783 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6l529" event={"ID":"3153dd6b-a60e-4da6-8463-a6e7b86c0277","Type":"ContainerStarted","Data":"af5a7c6287685abf347de70e2ea36c0c02768d9cb3992f09515dd1b3d66e1533"} Dec 17 08:37:59 crc kubenswrapper[4966]: I1217 08:37:59.459841 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.468668 4966 generic.go:334] "Generic (PLEG): container finished" podID="84d6c04f-29ad-41b5-a6b7-a8868cf02155" containerID="399e5d4c8d544e1fd09d3550c8c61046fd1485a2660334ccd4ae7f4d6f2e042c" exitCode=0 Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.468857 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0937-account-create-update-bq6tm" event={"ID":"84d6c04f-29ad-41b5-a6b7-a8868cf02155","Type":"ContainerDied","Data":"399e5d4c8d544e1fd09d3550c8c61046fd1485a2660334ccd4ae7f4d6f2e042c"} Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.470757 4966 generic.go:334] "Generic (PLEG): container finished" podID="3153dd6b-a60e-4da6-8463-a6e7b86c0277" containerID="76a94e47d47be924f17c05bfa6c88dbcf96fbbfe725c870f8d3fd62d97ba32ed" exitCode=0 Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.471310 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6l529" event={"ID":"3153dd6b-a60e-4da6-8463-a6e7b86c0277","Type":"ContainerDied","Data":"76a94e47d47be924f17c05bfa6c88dbcf96fbbfe725c870f8d3fd62d97ba32ed"} Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.596327 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:38:00 crc kubenswrapper[4966]: E1217 08:38:00.596894 4966 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 17 08:38:00 crc kubenswrapper[4966]: E1217 08:38:00.596923 4966 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 17 08:38:00 crc kubenswrapper[4966]: E1217 08:38:00.596972 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift podName:cbc21129-413d-4f88-9287-85c9a0fcfa68 nodeName:}" failed. No retries permitted until 2025-12-17 08:38:04.596955848 +0000 UTC m=+1020.142025790 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift") pod "swift-storage-0" (UID: "cbc21129-413d-4f88-9287-85c9a0fcfa68") : configmap "swift-ring-files" not found Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.641987 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-sbrnl"] Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.643375 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.647103 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.647344 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.647626 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.667655 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-sbrnl"] Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.698354 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-scripts\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.698442 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-ring-data-devices\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.698593 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjnv7\" (UniqueName: \"kubernetes.io/projected/f84813e1-82c5-456c-9dd9-ae14bed030db-kube-api-access-wjnv7\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.698677 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-dispersionconf\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.698717 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-swiftconf\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.698772 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-combined-ca-bundle\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.698898 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84813e1-82c5-456c-9dd9-ae14bed030db-etc-swift\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.800737 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-dispersionconf\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.800796 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-swiftconf\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.800829 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-combined-ca-bundle\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.801043 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84813e1-82c5-456c-9dd9-ae14bed030db-etc-swift\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.801672 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-scripts\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.801737 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-scripts\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.801759 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-ring-data-devices\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.801817 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjnv7\" (UniqueName: \"kubernetes.io/projected/f84813e1-82c5-456c-9dd9-ae14bed030db-kube-api-access-wjnv7\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.802272 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-ring-data-devices\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.802365 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84813e1-82c5-456c-9dd9-ae14bed030db-etc-swift\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.806000 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-dispersionconf\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.808469 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-swiftconf\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.811161 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-combined-ca-bundle\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.822403 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjnv7\" (UniqueName: \"kubernetes.io/projected/f84813e1-82c5-456c-9dd9-ae14bed030db-kube-api-access-wjnv7\") pod \"swift-ring-rebalance-sbrnl\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:00 crc kubenswrapper[4966]: I1217 08:38:00.965059 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:01 crc kubenswrapper[4966]: W1217 08:38:01.444351 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf84813e1_82c5_456c_9dd9_ae14bed030db.slice/crio-bcb91ce45b504b9bbb12dde95c18bb34e5cc60941dcbae11cd851181756b7fed WatchSource:0}: Error finding container bcb91ce45b504b9bbb12dde95c18bb34e5cc60941dcbae11cd851181756b7fed: Status 404 returned error can't find the container with id bcb91ce45b504b9bbb12dde95c18bb34e5cc60941dcbae11cd851181756b7fed Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.449612 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-sbrnl"] Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.478903 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sbrnl" event={"ID":"f84813e1-82c5-456c-9dd9-ae14bed030db","Type":"ContainerStarted","Data":"bcb91ce45b504b9bbb12dde95c18bb34e5cc60941dcbae11cd851181756b7fed"} Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.753346 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.826488 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d6c04f-29ad-41b5-a6b7-a8868cf02155-operator-scripts\") pod \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\" (UID: \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\") " Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.826563 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xxqj\" (UniqueName: \"kubernetes.io/projected/84d6c04f-29ad-41b5-a6b7-a8868cf02155-kube-api-access-8xxqj\") pod \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\" (UID: \"84d6c04f-29ad-41b5-a6b7-a8868cf02155\") " Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.828137 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d6c04f-29ad-41b5-a6b7-a8868cf02155-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84d6c04f-29ad-41b5-a6b7-a8868cf02155" (UID: "84d6c04f-29ad-41b5-a6b7-a8868cf02155"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.835053 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d6c04f-29ad-41b5-a6b7-a8868cf02155-kube-api-access-8xxqj" (OuterVolumeSpecName: "kube-api-access-8xxqj") pod "84d6c04f-29ad-41b5-a6b7-a8868cf02155" (UID: "84d6c04f-29ad-41b5-a6b7-a8868cf02155"). InnerVolumeSpecName "kube-api-access-8xxqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.918984 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6l529" Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.940666 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84d6c04f-29ad-41b5-a6b7-a8868cf02155-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:01 crc kubenswrapper[4966]: I1217 08:38:01.940980 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xxqj\" (UniqueName: \"kubernetes.io/projected/84d6c04f-29ad-41b5-a6b7-a8868cf02155-kube-api-access-8xxqj\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.041939 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csbm9\" (UniqueName: \"kubernetes.io/projected/3153dd6b-a60e-4da6-8463-a6e7b86c0277-kube-api-access-csbm9\") pod \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\" (UID: \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\") " Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.042042 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3153dd6b-a60e-4da6-8463-a6e7b86c0277-operator-scripts\") pod \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\" (UID: \"3153dd6b-a60e-4da6-8463-a6e7b86c0277\") " Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.042805 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3153dd6b-a60e-4da6-8463-a6e7b86c0277-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3153dd6b-a60e-4da6-8463-a6e7b86c0277" (UID: "3153dd6b-a60e-4da6-8463-a6e7b86c0277"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.047154 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3153dd6b-a60e-4da6-8463-a6e7b86c0277-kube-api-access-csbm9" (OuterVolumeSpecName: "kube-api-access-csbm9") pod "3153dd6b-a60e-4da6-8463-a6e7b86c0277" (UID: "3153dd6b-a60e-4da6-8463-a6e7b86c0277"). InnerVolumeSpecName "kube-api-access-csbm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.145036 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csbm9\" (UniqueName: \"kubernetes.io/projected/3153dd6b-a60e-4da6-8463-a6e7b86c0277-kube-api-access-csbm9\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.145090 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3153dd6b-a60e-4da6-8463-a6e7b86c0277-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.488212 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0937-account-create-update-bq6tm" event={"ID":"84d6c04f-29ad-41b5-a6b7-a8868cf02155","Type":"ContainerDied","Data":"aa22da15c1e0ac91e15497394f386426ea54867a58a3370e0517af07085aa49c"} Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.488255 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa22da15c1e0ac91e15497394f386426ea54867a58a3370e0517af07085aa49c" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.488320 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0937-account-create-update-bq6tm" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.501941 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6l529" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.501967 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6l529" event={"ID":"3153dd6b-a60e-4da6-8463-a6e7b86c0277","Type":"ContainerDied","Data":"af5a7c6287685abf347de70e2ea36c0c02768d9cb3992f09515dd1b3d66e1533"} Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.502007 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af5a7c6287685abf347de70e2ea36c0c02768d9cb3992f09515dd1b3d66e1533" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.741058 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-c68tj"] Dec 17 08:38:02 crc kubenswrapper[4966]: E1217 08:38:02.741539 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d6c04f-29ad-41b5-a6b7-a8868cf02155" containerName="mariadb-account-create-update" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.741559 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d6c04f-29ad-41b5-a6b7-a8868cf02155" containerName="mariadb-account-create-update" Dec 17 08:38:02 crc kubenswrapper[4966]: E1217 08:38:02.741589 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3153dd6b-a60e-4da6-8463-a6e7b86c0277" containerName="mariadb-database-create" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.741598 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3153dd6b-a60e-4da6-8463-a6e7b86c0277" containerName="mariadb-database-create" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.741837 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d6c04f-29ad-41b5-a6b7-a8868cf02155" containerName="mariadb-account-create-update" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.741858 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="3153dd6b-a60e-4da6-8463-a6e7b86c0277" containerName="mariadb-database-create" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.742588 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.747246 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-c68tj"] Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.859921 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gcbt\" (UniqueName: \"kubernetes.io/projected/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-kube-api-access-8gcbt\") pod \"keystone-db-create-c68tj\" (UID: \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\") " pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.860321 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-operator-scripts\") pod \"keystone-db-create-c68tj\" (UID: \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\") " pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.903080 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-382e-account-create-update-tvqcb"] Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.904861 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.906971 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.912337 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-382e-account-create-update-tvqcb"] Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.962008 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gcbt\" (UniqueName: \"kubernetes.io/projected/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-kube-api-access-8gcbt\") pod \"keystone-db-create-c68tj\" (UID: \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\") " pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.967024 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-operator-scripts\") pod \"keystone-db-create-c68tj\" (UID: \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\") " pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.968043 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-operator-scripts\") pod \"keystone-db-create-c68tj\" (UID: \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\") " pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:02 crc kubenswrapper[4966]: I1217 08:38:02.987609 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gcbt\" (UniqueName: \"kubernetes.io/projected/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-kube-api-access-8gcbt\") pod \"keystone-db-create-c68tj\" (UID: \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\") " pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.058437 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-znk6r"] Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.059947 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-znk6r" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.067773 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-znk6r"] Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.068453 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a603f92f-57a4-474a-b8e8-175f528a7a08-operator-scripts\") pod \"keystone-382e-account-create-update-tvqcb\" (UID: \"a603f92f-57a4-474a-b8e8-175f528a7a08\") " pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.068513 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwtsp\" (UniqueName: \"kubernetes.io/projected/a603f92f-57a4-474a-b8e8-175f528a7a08-kube-api-access-rwtsp\") pod \"keystone-382e-account-create-update-tvqcb\" (UID: \"a603f92f-57a4-474a-b8e8-175f528a7a08\") " pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.073128 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.166010 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-097c-account-create-update-sg2zd"] Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.166971 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.168923 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.169808 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a603f92f-57a4-474a-b8e8-175f528a7a08-operator-scripts\") pod \"keystone-382e-account-create-update-tvqcb\" (UID: \"a603f92f-57a4-474a-b8e8-175f528a7a08\") " pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.169944 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwtsp\" (UniqueName: \"kubernetes.io/projected/a603f92f-57a4-474a-b8e8-175f528a7a08-kube-api-access-rwtsp\") pod \"keystone-382e-account-create-update-tvqcb\" (UID: \"a603f92f-57a4-474a-b8e8-175f528a7a08\") " pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.170027 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24d6079f-863e-48f6-a8d8-66ffda331c89-operator-scripts\") pod \"placement-db-create-znk6r\" (UID: \"24d6079f-863e-48f6-a8d8-66ffda331c89\") " pod="openstack/placement-db-create-znk6r" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.170066 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz7pj\" (UniqueName: \"kubernetes.io/projected/24d6079f-863e-48f6-a8d8-66ffda331c89-kube-api-access-nz7pj\") pod \"placement-db-create-znk6r\" (UID: \"24d6079f-863e-48f6-a8d8-66ffda331c89\") " pod="openstack/placement-db-create-znk6r" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.170659 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a603f92f-57a4-474a-b8e8-175f528a7a08-operator-scripts\") pod \"keystone-382e-account-create-update-tvqcb\" (UID: \"a603f92f-57a4-474a-b8e8-175f528a7a08\") " pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.181972 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-097c-account-create-update-sg2zd"] Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.211832 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwtsp\" (UniqueName: \"kubernetes.io/projected/a603f92f-57a4-474a-b8e8-175f528a7a08-kube-api-access-rwtsp\") pod \"keystone-382e-account-create-update-tvqcb\" (UID: \"a603f92f-57a4-474a-b8e8-175f528a7a08\") " pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.238939 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.271765 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd2zh\" (UniqueName: \"kubernetes.io/projected/00699bc6-288b-4a56-92f4-280d1371f224-kube-api-access-fd2zh\") pod \"placement-097c-account-create-update-sg2zd\" (UID: \"00699bc6-288b-4a56-92f4-280d1371f224\") " pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.271915 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24d6079f-863e-48f6-a8d8-66ffda331c89-operator-scripts\") pod \"placement-db-create-znk6r\" (UID: \"24d6079f-863e-48f6-a8d8-66ffda331c89\") " pod="openstack/placement-db-create-znk6r" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.271955 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz7pj\" (UniqueName: \"kubernetes.io/projected/24d6079f-863e-48f6-a8d8-66ffda331c89-kube-api-access-nz7pj\") pod \"placement-db-create-znk6r\" (UID: \"24d6079f-863e-48f6-a8d8-66ffda331c89\") " pod="openstack/placement-db-create-znk6r" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.272042 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00699bc6-288b-4a56-92f4-280d1371f224-operator-scripts\") pod \"placement-097c-account-create-update-sg2zd\" (UID: \"00699bc6-288b-4a56-92f4-280d1371f224\") " pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.272977 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24d6079f-863e-48f6-a8d8-66ffda331c89-operator-scripts\") pod \"placement-db-create-znk6r\" (UID: \"24d6079f-863e-48f6-a8d8-66ffda331c89\") " pod="openstack/placement-db-create-znk6r" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.288769 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz7pj\" (UniqueName: \"kubernetes.io/projected/24d6079f-863e-48f6-a8d8-66ffda331c89-kube-api-access-nz7pj\") pod \"placement-db-create-znk6r\" (UID: \"24d6079f-863e-48f6-a8d8-66ffda331c89\") " pod="openstack/placement-db-create-znk6r" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.373894 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd2zh\" (UniqueName: \"kubernetes.io/projected/00699bc6-288b-4a56-92f4-280d1371f224-kube-api-access-fd2zh\") pod \"placement-097c-account-create-update-sg2zd\" (UID: \"00699bc6-288b-4a56-92f4-280d1371f224\") " pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.375936 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00699bc6-288b-4a56-92f4-280d1371f224-operator-scripts\") pod \"placement-097c-account-create-update-sg2zd\" (UID: \"00699bc6-288b-4a56-92f4-280d1371f224\") " pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.377561 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00699bc6-288b-4a56-92f4-280d1371f224-operator-scripts\") pod \"placement-097c-account-create-update-sg2zd\" (UID: \"00699bc6-288b-4a56-92f4-280d1371f224\") " pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.382414 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-znk6r" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.390652 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd2zh\" (UniqueName: \"kubernetes.io/projected/00699bc6-288b-4a56-92f4-280d1371f224-kube-api-access-fd2zh\") pod \"placement-097c-account-create-update-sg2zd\" (UID: \"00699bc6-288b-4a56-92f4-280d1371f224\") " pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.488075 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.644288 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-vp2ln"] Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.645254 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.649741 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.649778 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8vp7n" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.654589 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vp2ln"] Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.788858 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-combined-ca-bundle\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.788988 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-db-sync-config-data\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.789064 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-config-data\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.789126 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khtfj\" (UniqueName: \"kubernetes.io/projected/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-kube-api-access-khtfj\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.890368 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-combined-ca-bundle\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.890431 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-db-sync-config-data\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.890491 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-config-data\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.890573 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khtfj\" (UniqueName: \"kubernetes.io/projected/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-kube-api-access-khtfj\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.895651 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-db-sync-config-data\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.903339 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-combined-ca-bundle\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.903830 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-config-data\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.910649 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khtfj\" (UniqueName: \"kubernetes.io/projected/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-kube-api-access-khtfj\") pod \"glance-db-sync-vp2ln\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:03 crc kubenswrapper[4966]: I1217 08:38:03.968130 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:04 crc kubenswrapper[4966]: I1217 08:38:04.603731 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:38:04 crc kubenswrapper[4966]: E1217 08:38:04.604194 4966 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 17 08:38:04 crc kubenswrapper[4966]: E1217 08:38:04.604525 4966 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 17 08:38:04 crc kubenswrapper[4966]: E1217 08:38:04.604575 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift podName:cbc21129-413d-4f88-9287-85c9a0fcfa68 nodeName:}" failed. No retries permitted until 2025-12-17 08:38:12.604558745 +0000 UTC m=+1028.149628677 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift") pod "swift-storage-0" (UID: "cbc21129-413d-4f88-9287-85c9a0fcfa68") : configmap "swift-ring-files" not found Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.175366 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-znk6r"] Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.485932 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-c68tj"] Dec 17 08:38:05 crc kubenswrapper[4966]: W1217 08:38:05.528237 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda603f92f_57a4_474a_b8e8_175f528a7a08.slice/crio-e0bc2ddfcaa9cb672bb490a36df87318792294a5a6e04c0fbff80541b00352ea WatchSource:0}: Error finding container e0bc2ddfcaa9cb672bb490a36df87318792294a5a6e04c0fbff80541b00352ea: Status 404 returned error can't find the container with id e0bc2ddfcaa9cb672bb490a36df87318792294a5a6e04c0fbff80541b00352ea Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.536066 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sbrnl" event={"ID":"f84813e1-82c5-456c-9dd9-ae14bed030db","Type":"ContainerStarted","Data":"d5173eae20a3958a41ea3a86d8bd379203921e56d412c7f63afb5c98a9c03e7e"} Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.537042 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.539950 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-382e-account-create-update-tvqcb"] Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.540635 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-znk6r" event={"ID":"24d6079f-863e-48f6-a8d8-66ffda331c89","Type":"ContainerStarted","Data":"dd00a55a88218f1f7dc9298b73c2400bce6f4f0fc57cc40e25643031b577826b"} Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.540668 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-znk6r" event={"ID":"24d6079f-863e-48f6-a8d8-66ffda331c89","Type":"ContainerStarted","Data":"259ff108769e082691c712bf2fa5f41fb5b385fb6205b4893b72deb548617bf3"} Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.554680 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c68tj" event={"ID":"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183","Type":"ContainerStarted","Data":"0e1ce9d1f852b3488add4302fc4bdc24508f2af1f7d93e718ddb415e9581820a"} Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.559845 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-097c-account-create-update-sg2zd"] Dec 17 08:38:05 crc kubenswrapper[4966]: W1217 08:38:05.584218 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00699bc6_288b_4a56_92f4_280d1371f224.slice/crio-7e4889bd19de2bb3f93a4ed577cf78159538b9094886afce083340afca846a6f WatchSource:0}: Error finding container 7e4889bd19de2bb3f93a4ed577cf78159538b9094886afce083340afca846a6f: Status 404 returned error can't find the container with id 7e4889bd19de2bb3f93a4ed577cf78159538b9094886afce083340afca846a6f Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.589725 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-sbrnl" podStartSLOduration=2.3436439350000002 podStartE2EDuration="5.589704728s" podCreationTimestamp="2025-12-17 08:38:00 +0000 UTC" firstStartedPulling="2025-12-17 08:38:01.446662969 +0000 UTC m=+1016.991732911" lastFinishedPulling="2025-12-17 08:38:04.692723762 +0000 UTC m=+1020.237793704" observedRunningTime="2025-12-17 08:38:05.565504065 +0000 UTC m=+1021.110574017" watchObservedRunningTime="2025-12-17 08:38:05.589704728 +0000 UTC m=+1021.134774670" Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.600401 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-znk6r" podStartSLOduration=2.600382901 podStartE2EDuration="2.600382901s" podCreationTimestamp="2025-12-17 08:38:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:05.582702476 +0000 UTC m=+1021.127772418" watchObservedRunningTime="2025-12-17 08:38:05.600382901 +0000 UTC m=+1021.145452843" Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.603945 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.641181 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vp2ln"] Dec 17 08:38:05 crc kubenswrapper[4966]: I1217 08:38:05.943141 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.035395 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-647ff6c755-tzk8k"] Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.035936 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" podUID="3e81f86f-d072-4057-807d-81d017f072f9" containerName="dnsmasq-dns" containerID="cri-o://1ca1f40e30419dcae865d389041d39c16f770b817813aaaf6b8415cfc0e53fbf" gracePeriod=10 Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.565759 4966 generic.go:334] "Generic (PLEG): container finished" podID="24d6079f-863e-48f6-a8d8-66ffda331c89" containerID="dd00a55a88218f1f7dc9298b73c2400bce6f4f0fc57cc40e25643031b577826b" exitCode=0 Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.565837 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-znk6r" event={"ID":"24d6079f-863e-48f6-a8d8-66ffda331c89","Type":"ContainerDied","Data":"dd00a55a88218f1f7dc9298b73c2400bce6f4f0fc57cc40e25643031b577826b"} Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.567257 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vp2ln" event={"ID":"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06","Type":"ContainerStarted","Data":"4f059ee8e957ed0e3c118ec50c204493396b71615b51d5b5039a6158fbfba5d4"} Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.570695 4966 generic.go:334] "Generic (PLEG): container finished" podID="3e81f86f-d072-4057-807d-81d017f072f9" containerID="1ca1f40e30419dcae865d389041d39c16f770b817813aaaf6b8415cfc0e53fbf" exitCode=0 Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.570786 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" event={"ID":"3e81f86f-d072-4057-807d-81d017f072f9","Type":"ContainerDied","Data":"1ca1f40e30419dcae865d389041d39c16f770b817813aaaf6b8415cfc0e53fbf"} Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.570834 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" event={"ID":"3e81f86f-d072-4057-807d-81d017f072f9","Type":"ContainerDied","Data":"29b22ac1b59ac0d9501c81785144d921fcb3a4de838e9c9e85784f672b25c62a"} Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.570847 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29b22ac1b59ac0d9501c81785144d921fcb3a4de838e9c9e85784f672b25c62a" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.573041 4966 generic.go:334] "Generic (PLEG): container finished" podID="8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183" containerID="dcc7737790f864cddd18451c40f96b905d2b29ea5db9fa864afce06c80e6fa5d" exitCode=0 Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.573149 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c68tj" event={"ID":"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183","Type":"ContainerDied","Data":"dcc7737790f864cddd18451c40f96b905d2b29ea5db9fa864afce06c80e6fa5d"} Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.574854 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-382e-account-create-update-tvqcb" event={"ID":"a603f92f-57a4-474a-b8e8-175f528a7a08","Type":"ContainerStarted","Data":"24696c1cc38752d3b8fb935616ad75818333cf6e8043d3d2bf8a31fc814d5382"} Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.574908 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-382e-account-create-update-tvqcb" event={"ID":"a603f92f-57a4-474a-b8e8-175f528a7a08","Type":"ContainerStarted","Data":"e0bc2ddfcaa9cb672bb490a36df87318792294a5a6e04c0fbff80541b00352ea"} Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.577171 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-097c-account-create-update-sg2zd" event={"ID":"00699bc6-288b-4a56-92f4-280d1371f224","Type":"ContainerStarted","Data":"fd04fc0fe9a3d320c1a8971bf2f586b5158d04fd71966ab3ef1a5dcf81409e6b"} Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.577200 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-097c-account-create-update-sg2zd" event={"ID":"00699bc6-288b-4a56-92f4-280d1371f224","Type":"ContainerStarted","Data":"7e4889bd19de2bb3f93a4ed577cf78159538b9094886afce083340afca846a6f"} Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.583379 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.608910 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-097c-account-create-update-sg2zd" podStartSLOduration=3.608867613 podStartE2EDuration="3.608867613s" podCreationTimestamp="2025-12-17 08:38:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:06.606494398 +0000 UTC m=+1022.151564330" watchObservedRunningTime="2025-12-17 08:38:06.608867613 +0000 UTC m=+1022.153937555" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.677740 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-382e-account-create-update-tvqcb" podStartSLOduration=4.67771835 podStartE2EDuration="4.67771835s" podCreationTimestamp="2025-12-17 08:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:06.669742731 +0000 UTC m=+1022.214812683" watchObservedRunningTime="2025-12-17 08:38:06.67771835 +0000 UTC m=+1022.222788292" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.743541 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-dns-svc\") pod \"3e81f86f-d072-4057-807d-81d017f072f9\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.743633 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-config\") pod \"3e81f86f-d072-4057-807d-81d017f072f9\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.743789 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snd6c\" (UniqueName: \"kubernetes.io/projected/3e81f86f-d072-4057-807d-81d017f072f9-kube-api-access-snd6c\") pod \"3e81f86f-d072-4057-807d-81d017f072f9\" (UID: \"3e81f86f-d072-4057-807d-81d017f072f9\") " Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.757176 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e81f86f-d072-4057-807d-81d017f072f9-kube-api-access-snd6c" (OuterVolumeSpecName: "kube-api-access-snd6c") pod "3e81f86f-d072-4057-807d-81d017f072f9" (UID: "3e81f86f-d072-4057-807d-81d017f072f9"). InnerVolumeSpecName "kube-api-access-snd6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.788594 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-config" (OuterVolumeSpecName: "config") pod "3e81f86f-d072-4057-807d-81d017f072f9" (UID: "3e81f86f-d072-4057-807d-81d017f072f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.789460 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e81f86f-d072-4057-807d-81d017f072f9" (UID: "3e81f86f-d072-4057-807d-81d017f072f9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.845573 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snd6c\" (UniqueName: \"kubernetes.io/projected/3e81f86f-d072-4057-807d-81d017f072f9-kube-api-access-snd6c\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.845607 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.845617 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e81f86f-d072-4057-807d-81d017f072f9-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:06 crc kubenswrapper[4966]: I1217 08:38:06.930809 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.585392 4966 generic.go:334] "Generic (PLEG): container finished" podID="00699bc6-288b-4a56-92f4-280d1371f224" containerID="fd04fc0fe9a3d320c1a8971bf2f586b5158d04fd71966ab3ef1a5dcf81409e6b" exitCode=0 Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.585451 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-097c-account-create-update-sg2zd" event={"ID":"00699bc6-288b-4a56-92f4-280d1371f224","Type":"ContainerDied","Data":"fd04fc0fe9a3d320c1a8971bf2f586b5158d04fd71966ab3ef1a5dcf81409e6b"} Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.587749 4966 generic.go:334] "Generic (PLEG): container finished" podID="a603f92f-57a4-474a-b8e8-175f528a7a08" containerID="24696c1cc38752d3b8fb935616ad75818333cf6e8043d3d2bf8a31fc814d5382" exitCode=0 Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.588005 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-382e-account-create-update-tvqcb" event={"ID":"a603f92f-57a4-474a-b8e8-175f528a7a08","Type":"ContainerDied","Data":"24696c1cc38752d3b8fb935616ad75818333cf6e8043d3d2bf8a31fc814d5382"} Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.588673 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-647ff6c755-tzk8k" Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.651093 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-647ff6c755-tzk8k"] Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.658681 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-647ff6c755-tzk8k"] Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.895856 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.968167 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-operator-scripts\") pod \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\" (UID: \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\") " Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.968389 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gcbt\" (UniqueName: \"kubernetes.io/projected/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-kube-api-access-8gcbt\") pod \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\" (UID: \"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183\") " Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.970075 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183" (UID: "8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:07 crc kubenswrapper[4966]: I1217 08:38:07.975723 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-kube-api-access-8gcbt" (OuterVolumeSpecName: "kube-api-access-8gcbt") pod "8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183" (UID: "8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183"). InnerVolumeSpecName "kube-api-access-8gcbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.054349 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-znk6r" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.070690 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.070725 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gcbt\" (UniqueName: \"kubernetes.io/projected/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183-kube-api-access-8gcbt\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.171400 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24d6079f-863e-48f6-a8d8-66ffda331c89-operator-scripts\") pod \"24d6079f-863e-48f6-a8d8-66ffda331c89\" (UID: \"24d6079f-863e-48f6-a8d8-66ffda331c89\") " Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.171970 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz7pj\" (UniqueName: \"kubernetes.io/projected/24d6079f-863e-48f6-a8d8-66ffda331c89-kube-api-access-nz7pj\") pod \"24d6079f-863e-48f6-a8d8-66ffda331c89\" (UID: \"24d6079f-863e-48f6-a8d8-66ffda331c89\") " Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.172213 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24d6079f-863e-48f6-a8d8-66ffda331c89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "24d6079f-863e-48f6-a8d8-66ffda331c89" (UID: "24d6079f-863e-48f6-a8d8-66ffda331c89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.172531 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24d6079f-863e-48f6-a8d8-66ffda331c89-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.190572 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d6079f-863e-48f6-a8d8-66ffda331c89-kube-api-access-nz7pj" (OuterVolumeSpecName: "kube-api-access-nz7pj") pod "24d6079f-863e-48f6-a8d8-66ffda331c89" (UID: "24d6079f-863e-48f6-a8d8-66ffda331c89"). InnerVolumeSpecName "kube-api-access-nz7pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.274319 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz7pj\" (UniqueName: \"kubernetes.io/projected/24d6079f-863e-48f6-a8d8-66ffda331c89-kube-api-access-nz7pj\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.598583 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-znk6r" event={"ID":"24d6079f-863e-48f6-a8d8-66ffda331c89","Type":"ContainerDied","Data":"259ff108769e082691c712bf2fa5f41fb5b385fb6205b4893b72deb548617bf3"} Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.598615 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="259ff108769e082691c712bf2fa5f41fb5b385fb6205b4893b72deb548617bf3" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.598619 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-znk6r" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.613093 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-c68tj" event={"ID":"8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183","Type":"ContainerDied","Data":"0e1ce9d1f852b3488add4302fc4bdc24508f2af1f7d93e718ddb415e9581820a"} Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.613138 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-c68tj" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.613138 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e1ce9d1f852b3488add4302fc4bdc24508f2af1f7d93e718ddb415e9581820a" Dec 17 08:38:08 crc kubenswrapper[4966]: I1217 08:38:08.853979 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e81f86f-d072-4057-807d-81d017f072f9" path="/var/lib/kubelet/pods/3e81f86f-d072-4057-807d-81d017f072f9/volumes" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.018079 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.023015 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.093336 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwtsp\" (UniqueName: \"kubernetes.io/projected/a603f92f-57a4-474a-b8e8-175f528a7a08-kube-api-access-rwtsp\") pod \"a603f92f-57a4-474a-b8e8-175f528a7a08\" (UID: \"a603f92f-57a4-474a-b8e8-175f528a7a08\") " Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.093453 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd2zh\" (UniqueName: \"kubernetes.io/projected/00699bc6-288b-4a56-92f4-280d1371f224-kube-api-access-fd2zh\") pod \"00699bc6-288b-4a56-92f4-280d1371f224\" (UID: \"00699bc6-288b-4a56-92f4-280d1371f224\") " Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.093516 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a603f92f-57a4-474a-b8e8-175f528a7a08-operator-scripts\") pod \"a603f92f-57a4-474a-b8e8-175f528a7a08\" (UID: \"a603f92f-57a4-474a-b8e8-175f528a7a08\") " Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.093542 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00699bc6-288b-4a56-92f4-280d1371f224-operator-scripts\") pod \"00699bc6-288b-4a56-92f4-280d1371f224\" (UID: \"00699bc6-288b-4a56-92f4-280d1371f224\") " Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.094258 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00699bc6-288b-4a56-92f4-280d1371f224-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00699bc6-288b-4a56-92f4-280d1371f224" (UID: "00699bc6-288b-4a56-92f4-280d1371f224"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.094489 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a603f92f-57a4-474a-b8e8-175f528a7a08-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a603f92f-57a4-474a-b8e8-175f528a7a08" (UID: "a603f92f-57a4-474a-b8e8-175f528a7a08"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.106181 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a603f92f-57a4-474a-b8e8-175f528a7a08-kube-api-access-rwtsp" (OuterVolumeSpecName: "kube-api-access-rwtsp") pod "a603f92f-57a4-474a-b8e8-175f528a7a08" (UID: "a603f92f-57a4-474a-b8e8-175f528a7a08"). InnerVolumeSpecName "kube-api-access-rwtsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.111644 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00699bc6-288b-4a56-92f4-280d1371f224-kube-api-access-fd2zh" (OuterVolumeSpecName: "kube-api-access-fd2zh") pod "00699bc6-288b-4a56-92f4-280d1371f224" (UID: "00699bc6-288b-4a56-92f4-280d1371f224"). InnerVolumeSpecName "kube-api-access-fd2zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.196056 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a603f92f-57a4-474a-b8e8-175f528a7a08-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.196090 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00699bc6-288b-4a56-92f4-280d1371f224-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.196103 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwtsp\" (UniqueName: \"kubernetes.io/projected/a603f92f-57a4-474a-b8e8-175f528a7a08-kube-api-access-rwtsp\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.196113 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd2zh\" (UniqueName: \"kubernetes.io/projected/00699bc6-288b-4a56-92f4-280d1371f224-kube-api-access-fd2zh\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.624284 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-382e-account-create-update-tvqcb" event={"ID":"a603f92f-57a4-474a-b8e8-175f528a7a08","Type":"ContainerDied","Data":"e0bc2ddfcaa9cb672bb490a36df87318792294a5a6e04c0fbff80541b00352ea"} Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.624604 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0bc2ddfcaa9cb672bb490a36df87318792294a5a6e04c0fbff80541b00352ea" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.624312 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-382e-account-create-update-tvqcb" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.630032 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-097c-account-create-update-sg2zd" event={"ID":"00699bc6-288b-4a56-92f4-280d1371f224","Type":"ContainerDied","Data":"7e4889bd19de2bb3f93a4ed577cf78159538b9094886afce083340afca846a6f"} Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.630175 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e4889bd19de2bb3f93a4ed577cf78159538b9094886afce083340afca846a6f" Dec 17 08:38:09 crc kubenswrapper[4966]: I1217 08:38:09.630284 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-097c-account-create-update-sg2zd" Dec 17 08:38:12 crc kubenswrapper[4966]: I1217 08:38:12.660448 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:38:12 crc kubenswrapper[4966]: E1217 08:38:12.660666 4966 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 17 08:38:12 crc kubenswrapper[4966]: E1217 08:38:12.660811 4966 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 17 08:38:12 crc kubenswrapper[4966]: E1217 08:38:12.660904 4966 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift podName:cbc21129-413d-4f88-9287-85c9a0fcfa68 nodeName:}" failed. No retries permitted until 2025-12-17 08:38:28.660883967 +0000 UTC m=+1044.205953909 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift") pod "swift-storage-0" (UID: "cbc21129-413d-4f88-9287-85c9a0fcfa68") : configmap "swift-ring-files" not found Dec 17 08:38:13 crc kubenswrapper[4966]: I1217 08:38:13.734279 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-cbln8" podUID="3cb36f72-ce5e-4490-9009-7fe4443152b5" containerName="ovn-controller" probeResult="failure" output=< Dec 17 08:38:13 crc kubenswrapper[4966]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 17 08:38:13 crc kubenswrapper[4966]: > Dec 17 08:38:13 crc kubenswrapper[4966]: I1217 08:38:13.860800 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:38:13 crc kubenswrapper[4966]: I1217 08:38:13.867004 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-f6rhs" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.199661 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cbln8-config-7w2l9"] Dec 17 08:38:14 crc kubenswrapper[4966]: E1217 08:38:14.200066 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e81f86f-d072-4057-807d-81d017f072f9" containerName="dnsmasq-dns" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200084 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e81f86f-d072-4057-807d-81d017f072f9" containerName="dnsmasq-dns" Dec 17 08:38:14 crc kubenswrapper[4966]: E1217 08:38:14.200100 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00699bc6-288b-4a56-92f4-280d1371f224" containerName="mariadb-account-create-update" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200106 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="00699bc6-288b-4a56-92f4-280d1371f224" containerName="mariadb-account-create-update" Dec 17 08:38:14 crc kubenswrapper[4966]: E1217 08:38:14.200116 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183" containerName="mariadb-database-create" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200122 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183" containerName="mariadb-database-create" Dec 17 08:38:14 crc kubenswrapper[4966]: E1217 08:38:14.200132 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a603f92f-57a4-474a-b8e8-175f528a7a08" containerName="mariadb-account-create-update" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200138 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a603f92f-57a4-474a-b8e8-175f528a7a08" containerName="mariadb-account-create-update" Dec 17 08:38:14 crc kubenswrapper[4966]: E1217 08:38:14.200154 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d6079f-863e-48f6-a8d8-66ffda331c89" containerName="mariadb-database-create" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200159 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d6079f-863e-48f6-a8d8-66ffda331c89" containerName="mariadb-database-create" Dec 17 08:38:14 crc kubenswrapper[4966]: E1217 08:38:14.200169 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e81f86f-d072-4057-807d-81d017f072f9" containerName="init" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200175 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e81f86f-d072-4057-807d-81d017f072f9" containerName="init" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200317 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a603f92f-57a4-474a-b8e8-175f528a7a08" containerName="mariadb-account-create-update" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200327 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d6079f-863e-48f6-a8d8-66ffda331c89" containerName="mariadb-database-create" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200335 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183" containerName="mariadb-database-create" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200346 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="00699bc6-288b-4a56-92f4-280d1371f224" containerName="mariadb-account-create-update" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200362 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e81f86f-d072-4057-807d-81d017f072f9" containerName="dnsmasq-dns" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.200947 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.207173 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.260644 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cbln8-config-7w2l9"] Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.290770 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.290824 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-log-ovn\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.290882 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n28d6\" (UniqueName: \"kubernetes.io/projected/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-kube-api-access-n28d6\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.290913 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-scripts\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.290997 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run-ovn\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.291016 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-additional-scripts\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.392641 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run-ovn\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.392694 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-additional-scripts\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.392741 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.392760 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-log-ovn\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.392792 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n28d6\" (UniqueName: \"kubernetes.io/projected/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-kube-api-access-n28d6\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.392826 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-scripts\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.393155 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.393227 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run-ovn\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.393759 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-additional-scripts\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.393818 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-log-ovn\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.395033 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-scripts\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.471550 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n28d6\" (UniqueName: \"kubernetes.io/projected/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-kube-api-access-n28d6\") pod \"ovn-controller-cbln8-config-7w2l9\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.578989 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.671409 4966 generic.go:334] "Generic (PLEG): container finished" podID="1bd01548-3d59-4472-b8ef-a899b2cedee8" containerID="f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd" exitCode=0 Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.671626 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1bd01548-3d59-4472-b8ef-a899b2cedee8","Type":"ContainerDied","Data":"f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd"} Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.673298 4966 generic.go:334] "Generic (PLEG): container finished" podID="6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" containerID="40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a" exitCode=0 Dec 17 08:38:14 crc kubenswrapper[4966]: I1217 08:38:14.675518 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb","Type":"ContainerDied","Data":"40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a"} Dec 17 08:38:16 crc kubenswrapper[4966]: I1217 08:38:16.807443 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:38:16 crc kubenswrapper[4966]: I1217 08:38:16.807507 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:38:17 crc kubenswrapper[4966]: I1217 08:38:17.696814 4966 generic.go:334] "Generic (PLEG): container finished" podID="f84813e1-82c5-456c-9dd9-ae14bed030db" containerID="d5173eae20a3958a41ea3a86d8bd379203921e56d412c7f63afb5c98a9c03e7e" exitCode=0 Dec 17 08:38:17 crc kubenswrapper[4966]: I1217 08:38:17.696975 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sbrnl" event={"ID":"f84813e1-82c5-456c-9dd9-ae14bed030db","Type":"ContainerDied","Data":"d5173eae20a3958a41ea3a86d8bd379203921e56d412c7f63afb5c98a9c03e7e"} Dec 17 08:38:18 crc kubenswrapper[4966]: I1217 08:38:18.703092 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-cbln8" podUID="3cb36f72-ce5e-4490-9009-7fe4443152b5" containerName="ovn-controller" probeResult="failure" output=< Dec 17 08:38:18 crc kubenswrapper[4966]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 17 08:38:18 crc kubenswrapper[4966]: > Dec 17 08:38:21 crc kubenswrapper[4966]: E1217 08:38:21.238478 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-glance-api:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:38:21 crc kubenswrapper[4966]: E1217 08:38:21.238971 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-glance-api:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:38:21 crc kubenswrapper[4966]: E1217 08:38:21.239303 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-glance-api:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-khtfj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-vp2ln_openstack(99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:38:21 crc kubenswrapper[4966]: E1217 08:38:21.242111 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-vp2ln" podUID="99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.405271 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.548132 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-combined-ca-bundle\") pod \"f84813e1-82c5-456c-9dd9-ae14bed030db\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.548199 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-ring-data-devices\") pod \"f84813e1-82c5-456c-9dd9-ae14bed030db\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.548237 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-swiftconf\") pod \"f84813e1-82c5-456c-9dd9-ae14bed030db\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.548347 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-scripts\") pod \"f84813e1-82c5-456c-9dd9-ae14bed030db\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.548445 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84813e1-82c5-456c-9dd9-ae14bed030db-etc-swift\") pod \"f84813e1-82c5-456c-9dd9-ae14bed030db\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.548474 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-dispersionconf\") pod \"f84813e1-82c5-456c-9dd9-ae14bed030db\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.548536 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjnv7\" (UniqueName: \"kubernetes.io/projected/f84813e1-82c5-456c-9dd9-ae14bed030db-kube-api-access-wjnv7\") pod \"f84813e1-82c5-456c-9dd9-ae14bed030db\" (UID: \"f84813e1-82c5-456c-9dd9-ae14bed030db\") " Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.549438 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f84813e1-82c5-456c-9dd9-ae14bed030db" (UID: "f84813e1-82c5-456c-9dd9-ae14bed030db"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.549653 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f84813e1-82c5-456c-9dd9-ae14bed030db-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f84813e1-82c5-456c-9dd9-ae14bed030db" (UID: "f84813e1-82c5-456c-9dd9-ae14bed030db"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.553793 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f84813e1-82c5-456c-9dd9-ae14bed030db-kube-api-access-wjnv7" (OuterVolumeSpecName: "kube-api-access-wjnv7") pod "f84813e1-82c5-456c-9dd9-ae14bed030db" (UID: "f84813e1-82c5-456c-9dd9-ae14bed030db"). InnerVolumeSpecName "kube-api-access-wjnv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.560959 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f84813e1-82c5-456c-9dd9-ae14bed030db" (UID: "f84813e1-82c5-456c-9dd9-ae14bed030db"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.573278 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-scripts" (OuterVolumeSpecName: "scripts") pod "f84813e1-82c5-456c-9dd9-ae14bed030db" (UID: "f84813e1-82c5-456c-9dd9-ae14bed030db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.574042 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f84813e1-82c5-456c-9dd9-ae14bed030db" (UID: "f84813e1-82c5-456c-9dd9-ae14bed030db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.577535 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f84813e1-82c5-456c-9dd9-ae14bed030db" (UID: "f84813e1-82c5-456c-9dd9-ae14bed030db"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.633954 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cbln8-config-7w2l9"] Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.651814 4966 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.651919 4966 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.651936 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f84813e1-82c5-456c-9dd9-ae14bed030db-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.651948 4966 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f84813e1-82c5-456c-9dd9-ae14bed030db-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.651982 4966 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.652004 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjnv7\" (UniqueName: \"kubernetes.io/projected/f84813e1-82c5-456c-9dd9-ae14bed030db-kube-api-access-wjnv7\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.652020 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f84813e1-82c5-456c-9dd9-ae14bed030db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.735065 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cbln8-config-7w2l9" event={"ID":"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5","Type":"ContainerStarted","Data":"468fe9ade54afabd91e6750c93e5acdca169db675c7a052a90d6925dc74c2b7a"} Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.737064 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1bd01548-3d59-4472-b8ef-a899b2cedee8","Type":"ContainerStarted","Data":"c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90"} Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.737270 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.739269 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb","Type":"ContainerStarted","Data":"08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392"} Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.739527 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.740860 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sbrnl" event={"ID":"f84813e1-82c5-456c-9dd9-ae14bed030db","Type":"ContainerDied","Data":"bcb91ce45b504b9bbb12dde95c18bb34e5cc60941dcbae11cd851181756b7fed"} Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.740927 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcb91ce45b504b9bbb12dde95c18bb34e5cc60941dcbae11cd851181756b7fed" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.740908 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sbrnl" Dec 17 08:38:21 crc kubenswrapper[4966]: E1217 08:38:21.743738 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.156:5001/podified-antelope-centos9/openstack-glance-api:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/glance-db-sync-vp2ln" podUID="99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.780221 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=64.146234099 podStartE2EDuration="1m13.780207864s" podCreationTimestamp="2025-12-17 08:37:08 +0000 UTC" firstStartedPulling="2025-12-17 08:37:29.796958979 +0000 UTC m=+985.342028921" lastFinishedPulling="2025-12-17 08:37:39.430932734 +0000 UTC m=+994.976002686" observedRunningTime="2025-12-17 08:38:21.779005272 +0000 UTC m=+1037.324075214" watchObservedRunningTime="2025-12-17 08:38:21.780207864 +0000 UTC m=+1037.325277806" Dec 17 08:38:21 crc kubenswrapper[4966]: I1217 08:38:21.804744 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=63.43187901 podStartE2EDuration="1m13.804713766s" podCreationTimestamp="2025-12-17 08:37:08 +0000 UTC" firstStartedPulling="2025-12-17 08:37:30.296293816 +0000 UTC m=+985.841363758" lastFinishedPulling="2025-12-17 08:37:40.669128572 +0000 UTC m=+996.214198514" observedRunningTime="2025-12-17 08:38:21.801350185 +0000 UTC m=+1037.346420127" watchObservedRunningTime="2025-12-17 08:38:21.804713766 +0000 UTC m=+1037.349783698" Dec 17 08:38:22 crc kubenswrapper[4966]: I1217 08:38:22.749927 4966 generic.go:334] "Generic (PLEG): container finished" podID="9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" containerID="cad9a1835e206a3215c09f9b7051c5339008db162d3ffc20a032223d28376b8a" exitCode=0 Dec 17 08:38:22 crc kubenswrapper[4966]: I1217 08:38:22.750001 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cbln8-config-7w2l9" event={"ID":"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5","Type":"ContainerDied","Data":"cad9a1835e206a3215c09f9b7051c5339008db162d3ffc20a032223d28376b8a"} Dec 17 08:38:23 crc kubenswrapper[4966]: I1217 08:38:23.729165 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-cbln8" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.092101 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.201372 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-additional-scripts\") pod \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.201499 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run-ovn\") pod \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.201550 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-log-ovn\") pod \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.201665 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run\") pod \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.201712 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-scripts\") pod \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.201857 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n28d6\" (UniqueName: \"kubernetes.io/projected/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-kube-api-access-n28d6\") pod \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\" (UID: \"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5\") " Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.203355 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" (UID: "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.203483 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run" (OuterVolumeSpecName: "var-run") pod "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" (UID: "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.203546 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" (UID: "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.203860 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" (UID: "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.204221 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-scripts" (OuterVolumeSpecName: "scripts") pod "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" (UID: "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.223051 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-kube-api-access-n28d6" (OuterVolumeSpecName: "kube-api-access-n28d6") pod "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" (UID: "9dbef7cc-68d5-45f9-99e3-93ec980c2ec5"). InnerVolumeSpecName "kube-api-access-n28d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.304074 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n28d6\" (UniqueName: \"kubernetes.io/projected/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-kube-api-access-n28d6\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.304122 4966 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.304134 4966 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.304145 4966 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.304155 4966 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-var-run\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.304164 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.768273 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cbln8-config-7w2l9" event={"ID":"9dbef7cc-68d5-45f9-99e3-93ec980c2ec5","Type":"ContainerDied","Data":"468fe9ade54afabd91e6750c93e5acdca169db675c7a052a90d6925dc74c2b7a"} Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.768624 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="468fe9ade54afabd91e6750c93e5acdca169db675c7a052a90d6925dc74c2b7a" Dec 17 08:38:24 crc kubenswrapper[4966]: I1217 08:38:24.768471 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8-config-7w2l9" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.222628 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-cbln8-config-7w2l9"] Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.232101 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-cbln8-config-7w2l9"] Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.323730 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cbln8-config-xmjt7"] Dec 17 08:38:25 crc kubenswrapper[4966]: E1217 08:38:25.324129 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" containerName="ovn-config" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.324151 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" containerName="ovn-config" Dec 17 08:38:25 crc kubenswrapper[4966]: E1217 08:38:25.324165 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f84813e1-82c5-456c-9dd9-ae14bed030db" containerName="swift-ring-rebalance" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.324173 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f84813e1-82c5-456c-9dd9-ae14bed030db" containerName="swift-ring-rebalance" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.324333 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f84813e1-82c5-456c-9dd9-ae14bed030db" containerName="swift-ring-rebalance" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.324352 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" containerName="ovn-config" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.324853 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.327217 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.375066 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cbln8-config-xmjt7"] Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.429909 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7ljw\" (UniqueName: \"kubernetes.io/projected/79108c98-2b77-4335-86fd-a7d5f9c1d493-kube-api-access-z7ljw\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.430219 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-scripts\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.430393 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.430415 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run-ovn\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.430464 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-log-ovn\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.430603 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-additional-scripts\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.531676 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7ljw\" (UniqueName: \"kubernetes.io/projected/79108c98-2b77-4335-86fd-a7d5f9c1d493-kube-api-access-z7ljw\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.531727 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-scripts\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.531800 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.531819 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run-ovn\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.531845 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-log-ovn\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.531925 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-additional-scripts\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.532601 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-additional-scripts\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.534241 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.534263 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run-ovn\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.534341 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-log-ovn\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.534434 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-scripts\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.554626 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7ljw\" (UniqueName: \"kubernetes.io/projected/79108c98-2b77-4335-86fd-a7d5f9c1d493-kube-api-access-z7ljw\") pod \"ovn-controller-cbln8-config-xmjt7\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:25 crc kubenswrapper[4966]: I1217 08:38:25.644025 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:26 crc kubenswrapper[4966]: I1217 08:38:26.101311 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cbln8-config-xmjt7"] Dec 17 08:38:26 crc kubenswrapper[4966]: I1217 08:38:26.792789 4966 generic.go:334] "Generic (PLEG): container finished" podID="79108c98-2b77-4335-86fd-a7d5f9c1d493" containerID="8896caf4d77c1bdce82efba5b0a7f386f59c73c52fb327d369cc3355889b6597" exitCode=0 Dec 17 08:38:26 crc kubenswrapper[4966]: I1217 08:38:26.792850 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cbln8-config-xmjt7" event={"ID":"79108c98-2b77-4335-86fd-a7d5f9c1d493","Type":"ContainerDied","Data":"8896caf4d77c1bdce82efba5b0a7f386f59c73c52fb327d369cc3355889b6597"} Dec 17 08:38:26 crc kubenswrapper[4966]: I1217 08:38:26.793135 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cbln8-config-xmjt7" event={"ID":"79108c98-2b77-4335-86fd-a7d5f9c1d493","Type":"ContainerStarted","Data":"64de0896b36aef9422282f8c9e4db1d6f8872ac7deb2092c14fa1c7dda3ab8d2"} Dec 17 08:38:26 crc kubenswrapper[4966]: I1217 08:38:26.840679 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dbef7cc-68d5-45f9-99e3-93ec980c2ec5" path="/var/lib/kubelet/pods/9dbef7cc-68d5-45f9-99e3-93ec980c2ec5/volumes" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.243817 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278315 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-log-ovn\") pod \"79108c98-2b77-4335-86fd-a7d5f9c1d493\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278383 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7ljw\" (UniqueName: \"kubernetes.io/projected/79108c98-2b77-4335-86fd-a7d5f9c1d493-kube-api-access-z7ljw\") pod \"79108c98-2b77-4335-86fd-a7d5f9c1d493\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278428 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run\") pod \"79108c98-2b77-4335-86fd-a7d5f9c1d493\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278436 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "79108c98-2b77-4335-86fd-a7d5f9c1d493" (UID: "79108c98-2b77-4335-86fd-a7d5f9c1d493"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278548 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run" (OuterVolumeSpecName: "var-run") pod "79108c98-2b77-4335-86fd-a7d5f9c1d493" (UID: "79108c98-2b77-4335-86fd-a7d5f9c1d493"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278591 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-additional-scripts\") pod \"79108c98-2b77-4335-86fd-a7d5f9c1d493\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278615 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-scripts\") pod \"79108c98-2b77-4335-86fd-a7d5f9c1d493\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278645 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run-ovn\") pod \"79108c98-2b77-4335-86fd-a7d5f9c1d493\" (UID: \"79108c98-2b77-4335-86fd-a7d5f9c1d493\") " Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278966 4966 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278981 4966 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.278981 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "79108c98-2b77-4335-86fd-a7d5f9c1d493" (UID: "79108c98-2b77-4335-86fd-a7d5f9c1d493"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.279489 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "79108c98-2b77-4335-86fd-a7d5f9c1d493" (UID: "79108c98-2b77-4335-86fd-a7d5f9c1d493"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.279721 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-scripts" (OuterVolumeSpecName: "scripts") pod "79108c98-2b77-4335-86fd-a7d5f9c1d493" (UID: "79108c98-2b77-4335-86fd-a7d5f9c1d493"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.292249 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79108c98-2b77-4335-86fd-a7d5f9c1d493-kube-api-access-z7ljw" (OuterVolumeSpecName: "kube-api-access-z7ljw") pod "79108c98-2b77-4335-86fd-a7d5f9c1d493" (UID: "79108c98-2b77-4335-86fd-a7d5f9c1d493"). InnerVolumeSpecName "kube-api-access-z7ljw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.380447 4966 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.380488 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79108c98-2b77-4335-86fd-a7d5f9c1d493-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.380497 4966 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79108c98-2b77-4335-86fd-a7d5f9c1d493-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.380505 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7ljw\" (UniqueName: \"kubernetes.io/projected/79108c98-2b77-4335-86fd-a7d5f9c1d493-kube-api-access-z7ljw\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.685904 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.690603 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cbc21129-413d-4f88-9287-85c9a0fcfa68-etc-swift\") pod \"swift-storage-0\" (UID: \"cbc21129-413d-4f88-9287-85c9a0fcfa68\") " pod="openstack/swift-storage-0" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.806417 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cbln8-config-xmjt7" event={"ID":"79108c98-2b77-4335-86fd-a7d5f9c1d493","Type":"ContainerDied","Data":"64de0896b36aef9422282f8c9e4db1d6f8872ac7deb2092c14fa1c7dda3ab8d2"} Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.806620 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64de0896b36aef9422282f8c9e4db1d6f8872ac7deb2092c14fa1c7dda3ab8d2" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.806487 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cbln8-config-xmjt7" Dec 17 08:38:28 crc kubenswrapper[4966]: I1217 08:38:28.924049 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 17 08:38:29 crc kubenswrapper[4966]: I1217 08:38:29.322226 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-cbln8-config-xmjt7"] Dec 17 08:38:29 crc kubenswrapper[4966]: I1217 08:38:29.329938 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-cbln8-config-xmjt7"] Dec 17 08:38:29 crc kubenswrapper[4966]: I1217 08:38:29.488481 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 17 08:38:29 crc kubenswrapper[4966]: I1217 08:38:29.820481 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"eeb093b66a6873a3647f851eba066a6b84efb846dae8c50f8421e8e36d9012c1"} Dec 17 08:38:30 crc kubenswrapper[4966]: I1217 08:38:30.843077 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79108c98-2b77-4335-86fd-a7d5f9c1d493" path="/var/lib/kubelet/pods/79108c98-2b77-4335-86fd-a7d5f9c1d493/volumes" Dec 17 08:38:30 crc kubenswrapper[4966]: I1217 08:38:30.844101 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"ada5e3006dff05d71f0e44df023786068ccca2b590141ddb8a24e73881b3b3b3"} Dec 17 08:38:30 crc kubenswrapper[4966]: I1217 08:38:30.844127 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"56744b9b6d79983af8d48af0226578b0c5b1542d4a3ceb1c15aafa3ccef89e6d"} Dec 17 08:38:30 crc kubenswrapper[4966]: I1217 08:38:30.844136 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"85d1fa1e8b9ce6bcff03d975f896773731bd3adba4c72d9f4051b6af1c6fce8e"} Dec 17 08:38:31 crc kubenswrapper[4966]: I1217 08:38:31.844421 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"0c6636b810fd84268f158979f26975d20211b6352dce1307a18bccad1ab5fc5f"} Dec 17 08:38:32 crc kubenswrapper[4966]: I1217 08:38:32.867232 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"9a4525c213ca6801044be365cf90a5cf8d1f7d09e9eb539c3bfd5cab12bfd62c"} Dec 17 08:38:32 crc kubenswrapper[4966]: I1217 08:38:32.867602 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"9882609234f166d23b612a599d012bf37bdb06421cb6633b58fc70542ce34a4e"} Dec 17 08:38:32 crc kubenswrapper[4966]: I1217 08:38:32.867614 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"f1209021d7bd4f16acaee142a6bf51251fb46df9d3c7c6dd2259ae2ecf77543d"} Dec 17 08:38:32 crc kubenswrapper[4966]: I1217 08:38:32.867627 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"02bd659c4609d29f43e804c8d26b783951a71c0668856080d30f16ab2a3c0417"} Dec 17 08:38:33 crc kubenswrapper[4966]: I1217 08:38:33.879672 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"4f2e941cca7fd2c66a65df7a5b5d551e7036d7b362fc5291447d34e5c32bea84"} Dec 17 08:38:33 crc kubenswrapper[4966]: I1217 08:38:33.879983 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"e6ffabc153c5a21409d2499165ed73b3d39b4e1ceeef088590c6476e6d860501"} Dec 17 08:38:33 crc kubenswrapper[4966]: I1217 08:38:33.879994 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"a36db7c33604570e9003c3fa39362d9ae3f86d6342f3f9e97a56262ffc19e383"} Dec 17 08:38:34 crc kubenswrapper[4966]: I1217 08:38:34.898176 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"ea6fd64045f32db59d5b1a4c23f2c9136ccec9fba1c64a24a85287077a4d4226"} Dec 17 08:38:34 crc kubenswrapper[4966]: I1217 08:38:34.898218 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"d76de81a6a7f19152a8bd82170919d6a1f269eefea5348be1b76c43606a48115"} Dec 17 08:38:34 crc kubenswrapper[4966]: I1217 08:38:34.898228 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"b7393469d141cb24660d1519d1d4da3acbb9204b30d66e9ca97810d96acb6687"} Dec 17 08:38:34 crc kubenswrapper[4966]: I1217 08:38:34.898237 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cbc21129-413d-4f88-9287-85c9a0fcfa68","Type":"ContainerStarted","Data":"ccfecf41cf5a46505fd55d0cb129fe642437662ab9cb0c012a764e7b13852a7e"} Dec 17 08:38:34 crc kubenswrapper[4966]: I1217 08:38:34.901084 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vp2ln" event={"ID":"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06","Type":"ContainerStarted","Data":"730743ecd154551f18d4f8797663937b0d35e8a91fb5e32a07c62d2181f4c277"} Dec 17 08:38:34 crc kubenswrapper[4966]: I1217 08:38:34.943110 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.278520892 podStartE2EDuration="39.943093046s" podCreationTimestamp="2025-12-17 08:37:55 +0000 UTC" firstStartedPulling="2025-12-17 08:38:29.503593191 +0000 UTC m=+1045.048663133" lastFinishedPulling="2025-12-17 08:38:33.168165345 +0000 UTC m=+1048.713235287" observedRunningTime="2025-12-17 08:38:34.940742281 +0000 UTC m=+1050.485812233" watchObservedRunningTime="2025-12-17 08:38:34.943093046 +0000 UTC m=+1050.488162988" Dec 17 08:38:34 crc kubenswrapper[4966]: I1217 08:38:34.973892 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-vp2ln" podStartSLOduration=3.717280656 podStartE2EDuration="31.973852739s" podCreationTimestamp="2025-12-17 08:38:03 +0000 UTC" firstStartedPulling="2025-12-17 08:38:05.64996067 +0000 UTC m=+1021.195030612" lastFinishedPulling="2025-12-17 08:38:33.906532753 +0000 UTC m=+1049.451602695" observedRunningTime="2025-12-17 08:38:34.969311914 +0000 UTC m=+1050.514381856" watchObservedRunningTime="2025-12-17 08:38:34.973852739 +0000 UTC m=+1050.518922691" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.257044 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58d6b88575-cwxzl"] Dec 17 08:38:35 crc kubenswrapper[4966]: E1217 08:38:35.257646 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79108c98-2b77-4335-86fd-a7d5f9c1d493" containerName="ovn-config" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.257663 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="79108c98-2b77-4335-86fd-a7d5f9c1d493" containerName="ovn-config" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.257807 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="79108c98-2b77-4335-86fd-a7d5f9c1d493" containerName="ovn-config" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.258620 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.260361 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.274085 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58d6b88575-cwxzl"] Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.305550 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-sb\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.305591 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-nb\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.305670 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f7pc\" (UniqueName: \"kubernetes.io/projected/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-kube-api-access-5f7pc\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.305737 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-svc\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.305766 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-swift-storage-0\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.305848 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-config\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.409134 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f7pc\" (UniqueName: \"kubernetes.io/projected/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-kube-api-access-5f7pc\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.409535 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-svc\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.409685 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-swift-storage-0\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.409843 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-config\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.410027 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-nb\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.410125 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-sb\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.410631 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-svc\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.411110 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-sb\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.411227 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-swift-storage-0\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.411231 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-nb\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.411227 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-config\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.426660 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f7pc\" (UniqueName: \"kubernetes.io/projected/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-kube-api-access-5f7pc\") pod \"dnsmasq-dns-58d6b88575-cwxzl\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:35 crc kubenswrapper[4966]: I1217 08:38:35.577540 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:36 crc kubenswrapper[4966]: I1217 08:38:36.091073 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58d6b88575-cwxzl"] Dec 17 08:38:36 crc kubenswrapper[4966]: I1217 08:38:36.925861 4966 generic.go:334] "Generic (PLEG): container finished" podID="8c41e1fa-cfd8-466e-b775-d500b0cc4f54" containerID="7c97bb4b153934cc04531a97e03bc2a3177f9b31f75a0e8e6f7e978e4dbcf97f" exitCode=0 Dec 17 08:38:36 crc kubenswrapper[4966]: I1217 08:38:36.926084 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" event={"ID":"8c41e1fa-cfd8-466e-b775-d500b0cc4f54","Type":"ContainerDied","Data":"7c97bb4b153934cc04531a97e03bc2a3177f9b31f75a0e8e6f7e978e4dbcf97f"} Dec 17 08:38:36 crc kubenswrapper[4966]: I1217 08:38:36.926107 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" event={"ID":"8c41e1fa-cfd8-466e-b775-d500b0cc4f54","Type":"ContainerStarted","Data":"431ca7587cf62aa75ca02ce1a569cb1667f49a3391d891bc590cf1212e881ddd"} Dec 17 08:38:37 crc kubenswrapper[4966]: I1217 08:38:37.940126 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" event={"ID":"8c41e1fa-cfd8-466e-b775-d500b0cc4f54","Type":"ContainerStarted","Data":"75e21cd23cff7c9ef9f9ce1502b45476e56ea6e1655c148b2917ff3c8014ad85"} Dec 17 08:38:37 crc kubenswrapper[4966]: I1217 08:38:37.941601 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:37 crc kubenswrapper[4966]: I1217 08:38:37.975447 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" podStartSLOduration=2.97543218 podStartE2EDuration="2.97543218s" podCreationTimestamp="2025-12-17 08:38:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:37.972441819 +0000 UTC m=+1053.517511781" watchObservedRunningTime="2025-12-17 08:38:37.97543218 +0000 UTC m=+1053.520502122" Dec 17 08:38:39 crc kubenswrapper[4966]: I1217 08:38:39.538453 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:38:39 crc kubenswrapper[4966]: I1217 08:38:39.913684 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 17 08:38:40 crc kubenswrapper[4966]: I1217 08:38:40.972815 4966 generic.go:334] "Generic (PLEG): container finished" podID="99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" containerID="730743ecd154551f18d4f8797663937b0d35e8a91fb5e32a07c62d2181f4c277" exitCode=0 Dec 17 08:38:40 crc kubenswrapper[4966]: I1217 08:38:40.972903 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vp2ln" event={"ID":"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06","Type":"ContainerDied","Data":"730743ecd154551f18d4f8797663937b0d35e8a91fb5e32a07c62d2181f4c277"} Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.291172 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-jpx62"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.292332 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-jpx62" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.310496 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-jpx62"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.337091 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23a69382-8e08-47cf-a9aa-bc254229aa21-operator-scripts\") pod \"heat-db-create-jpx62\" (UID: \"23a69382-8e08-47cf-a9aa-bc254229aa21\") " pod="openstack/heat-db-create-jpx62" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.337172 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwgvt\" (UniqueName: \"kubernetes.io/projected/23a69382-8e08-47cf-a9aa-bc254229aa21-kube-api-access-dwgvt\") pod \"heat-db-create-jpx62\" (UID: \"23a69382-8e08-47cf-a9aa-bc254229aa21\") " pod="openstack/heat-db-create-jpx62" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.439113 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23a69382-8e08-47cf-a9aa-bc254229aa21-operator-scripts\") pod \"heat-db-create-jpx62\" (UID: \"23a69382-8e08-47cf-a9aa-bc254229aa21\") " pod="openstack/heat-db-create-jpx62" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.439186 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwgvt\" (UniqueName: \"kubernetes.io/projected/23a69382-8e08-47cf-a9aa-bc254229aa21-kube-api-access-dwgvt\") pod \"heat-db-create-jpx62\" (UID: \"23a69382-8e08-47cf-a9aa-bc254229aa21\") " pod="openstack/heat-db-create-jpx62" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.440553 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23a69382-8e08-47cf-a9aa-bc254229aa21-operator-scripts\") pod \"heat-db-create-jpx62\" (UID: \"23a69382-8e08-47cf-a9aa-bc254229aa21\") " pod="openstack/heat-db-create-jpx62" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.461480 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwgvt\" (UniqueName: \"kubernetes.io/projected/23a69382-8e08-47cf-a9aa-bc254229aa21-kube-api-access-dwgvt\") pod \"heat-db-create-jpx62\" (UID: \"23a69382-8e08-47cf-a9aa-bc254229aa21\") " pod="openstack/heat-db-create-jpx62" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.518697 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-f803-account-create-update-p88mq"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.519689 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.525211 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.538993 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-fp4tg"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.540121 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.545674 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f803-account-create-update-p88mq"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.606188 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-jpx62" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.608089 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-fp4tg"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.623236 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-e584-account-create-update-hk75s"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.624260 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.637116 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-e584-account-create-update-hk75s"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.643542 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.643681 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv7sg\" (UniqueName: \"kubernetes.io/projected/552f3591-2913-4bbb-bf5b-79ea4ba40778-kube-api-access-rv7sg\") pod \"barbican-f803-account-create-update-p88mq\" (UID: \"552f3591-2913-4bbb-bf5b-79ea4ba40778\") " pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.643798 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck6kz\" (UniqueName: \"kubernetes.io/projected/a952c157-89d5-4bb4-a624-b5b43babd9e0-kube-api-access-ck6kz\") pod \"barbican-db-create-fp4tg\" (UID: \"a952c157-89d5-4bb4-a624-b5b43babd9e0\") " pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.643824 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a952c157-89d5-4bb4-a624-b5b43babd9e0-operator-scripts\") pod \"barbican-db-create-fp4tg\" (UID: \"a952c157-89d5-4bb4-a624-b5b43babd9e0\") " pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.643887 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552f3591-2913-4bbb-bf5b-79ea4ba40778-operator-scripts\") pod \"barbican-f803-account-create-update-p88mq\" (UID: \"552f3591-2913-4bbb-bf5b-79ea4ba40778\") " pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.715196 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-rjxqq"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.723919 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.746849 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70fc4b4-6cf8-4852-827a-d2939c1b656a-operator-scripts\") pod \"heat-e584-account-create-update-hk75s\" (UID: \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\") " pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.746954 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck6kz\" (UniqueName: \"kubernetes.io/projected/a952c157-89d5-4bb4-a624-b5b43babd9e0-kube-api-access-ck6kz\") pod \"barbican-db-create-fp4tg\" (UID: \"a952c157-89d5-4bb4-a624-b5b43babd9e0\") " pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.746991 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a952c157-89d5-4bb4-a624-b5b43babd9e0-operator-scripts\") pod \"barbican-db-create-fp4tg\" (UID: \"a952c157-89d5-4bb4-a624-b5b43babd9e0\") " pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.747048 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552f3591-2913-4bbb-bf5b-79ea4ba40778-operator-scripts\") pod \"barbican-f803-account-create-update-p88mq\" (UID: \"552f3591-2913-4bbb-bf5b-79ea4ba40778\") " pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.747072 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spxqp\" (UniqueName: \"kubernetes.io/projected/f70fc4b4-6cf8-4852-827a-d2939c1b656a-kube-api-access-spxqp\") pod \"heat-e584-account-create-update-hk75s\" (UID: \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\") " pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.747177 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv7sg\" (UniqueName: \"kubernetes.io/projected/552f3591-2913-4bbb-bf5b-79ea4ba40778-kube-api-access-rv7sg\") pod \"barbican-f803-account-create-update-p88mq\" (UID: \"552f3591-2913-4bbb-bf5b-79ea4ba40778\") " pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.748342 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a952c157-89d5-4bb4-a624-b5b43babd9e0-operator-scripts\") pod \"barbican-db-create-fp4tg\" (UID: \"a952c157-89d5-4bb4-a624-b5b43babd9e0\") " pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.751095 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552f3591-2913-4bbb-bf5b-79ea4ba40778-operator-scripts\") pod \"barbican-f803-account-create-update-p88mq\" (UID: \"552f3591-2913-4bbb-bf5b-79ea4ba40778\") " pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.765633 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rjxqq"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.814430 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-s8bmz"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.815556 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.848850 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70fc4b4-6cf8-4852-827a-d2939c1b656a-operator-scripts\") pod \"heat-e584-account-create-update-hk75s\" (UID: \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\") " pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.849111 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv7sg\" (UniqueName: \"kubernetes.io/projected/552f3591-2913-4bbb-bf5b-79ea4ba40778-kube-api-access-rv7sg\") pod \"barbican-f803-account-create-update-p88mq\" (UID: \"552f3591-2913-4bbb-bf5b-79ea4ba40778\") " pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.849155 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9jj6\" (UniqueName: \"kubernetes.io/projected/be459859-4e2c-4548-a544-39bf15a731b0-kube-api-access-w9jj6\") pod \"cinder-db-create-rjxqq\" (UID: \"be459859-4e2c-4548-a544-39bf15a731b0\") " pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.849212 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spxqp\" (UniqueName: \"kubernetes.io/projected/f70fc4b4-6cf8-4852-827a-d2939c1b656a-kube-api-access-spxqp\") pod \"heat-e584-account-create-update-hk75s\" (UID: \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\") " pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.849234 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be459859-4e2c-4548-a544-39bf15a731b0-operator-scripts\") pod \"cinder-db-create-rjxqq\" (UID: \"be459859-4e2c-4548-a544-39bf15a731b0\") " pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.849485 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.849699 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck6kz\" (UniqueName: \"kubernetes.io/projected/a952c157-89d5-4bb4-a624-b5b43babd9e0-kube-api-access-ck6kz\") pod \"barbican-db-create-fp4tg\" (UID: \"a952c157-89d5-4bb4-a624-b5b43babd9e0\") " pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.849801 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.849991 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.850094 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-czvn7" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.852030 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.854559 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70fc4b4-6cf8-4852-827a-d2939c1b656a-operator-scripts\") pod \"heat-e584-account-create-update-hk75s\" (UID: \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\") " pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.860374 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.901422 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-s8bmz"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.902297 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spxqp\" (UniqueName: \"kubernetes.io/projected/f70fc4b4-6cf8-4852-827a-d2939c1b656a-kube-api-access-spxqp\") pod \"heat-e584-account-create-update-hk75s\" (UID: \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\") " pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.922638 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-952b-account-create-update-p5444"] Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.923722 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.931951 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.959824 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be459859-4e2c-4548-a544-39bf15a731b0-operator-scripts\") pod \"cinder-db-create-rjxqq\" (UID: \"be459859-4e2c-4548-a544-39bf15a731b0\") " pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.959892 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-combined-ca-bundle\") pod \"keystone-db-sync-s8bmz\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.960010 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxw99\" (UniqueName: \"kubernetes.io/projected/8603b503-2df3-4a05-84e7-024e63d2b455-kube-api-access-hxw99\") pod \"keystone-db-sync-s8bmz\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.960053 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-config-data\") pod \"keystone-db-sync-s8bmz\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.960113 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9jj6\" (UniqueName: \"kubernetes.io/projected/be459859-4e2c-4548-a544-39bf15a731b0-kube-api-access-w9jj6\") pod \"cinder-db-create-rjxqq\" (UID: \"be459859-4e2c-4548-a544-39bf15a731b0\") " pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:41 crc kubenswrapper[4966]: I1217 08:38:41.961235 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be459859-4e2c-4548-a544-39bf15a731b0-operator-scripts\") pod \"cinder-db-create-rjxqq\" (UID: \"be459859-4e2c-4548-a544-39bf15a731b0\") " pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:41.996749 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9jj6\" (UniqueName: \"kubernetes.io/projected/be459859-4e2c-4548-a544-39bf15a731b0-kube-api-access-w9jj6\") pod \"cinder-db-create-rjxqq\" (UID: \"be459859-4e2c-4548-a544-39bf15a731b0\") " pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.019353 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.022962 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-952b-account-create-update-p5444"] Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.051924 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-z57pq"] Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.053108 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.055859 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.061169 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/771eea8c-0e4e-471c-be4b-3f18e3d91adb-operator-scripts\") pod \"cinder-952b-account-create-update-p5444\" (UID: \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\") " pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.061237 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-combined-ca-bundle\") pod \"keystone-db-sync-s8bmz\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.061268 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bb8x\" (UniqueName: \"kubernetes.io/projected/771eea8c-0e4e-471c-be4b-3f18e3d91adb-kube-api-access-6bb8x\") pod \"cinder-952b-account-create-update-p5444\" (UID: \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\") " pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.061331 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxw99\" (UniqueName: \"kubernetes.io/projected/8603b503-2df3-4a05-84e7-024e63d2b455-kube-api-access-hxw99\") pod \"keystone-db-sync-s8bmz\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.061357 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-config-data\") pod \"keystone-db-sync-s8bmz\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.073462 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-z57pq"] Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.073743 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-combined-ca-bundle\") pod \"keystone-db-sync-s8bmz\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.080601 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-config-data\") pod \"keystone-db-sync-s8bmz\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.088439 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxw99\" (UniqueName: \"kubernetes.io/projected/8603b503-2df3-4a05-84e7-024e63d2b455-kube-api-access-hxw99\") pod \"keystone-db-sync-s8bmz\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.145675 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d3cd-account-create-update-4ccdt"] Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.147052 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.151435 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.164072 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/771eea8c-0e4e-471c-be4b-3f18e3d91adb-operator-scripts\") pod \"cinder-952b-account-create-update-p5444\" (UID: \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\") " pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.164793 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/771eea8c-0e4e-471c-be4b-3f18e3d91adb-operator-scripts\") pod \"cinder-952b-account-create-update-p5444\" (UID: \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\") " pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.164846 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-operator-scripts\") pod \"neutron-db-create-z57pq\" (UID: \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\") " pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.164982 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bb8x\" (UniqueName: \"kubernetes.io/projected/771eea8c-0e4e-471c-be4b-3f18e3d91adb-kube-api-access-6bb8x\") pod \"cinder-952b-account-create-update-p5444\" (UID: \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\") " pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.165368 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p642p\" (UniqueName: \"kubernetes.io/projected/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-kube-api-access-p642p\") pod \"neutron-db-create-z57pq\" (UID: \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\") " pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.167140 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d3cd-account-create-update-4ccdt"] Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.203472 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bb8x\" (UniqueName: \"kubernetes.io/projected/771eea8c-0e4e-471c-be4b-3f18e3d91adb-kube-api-access-6bb8x\") pod \"cinder-952b-account-create-update-p5444\" (UID: \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\") " pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.235305 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.250940 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.266837 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-operator-scripts\") pod \"neutron-db-create-z57pq\" (UID: \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\") " pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.266953 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937dc9ee-a952-4496-9659-b4cc9299018a-operator-scripts\") pod \"neutron-d3cd-account-create-update-4ccdt\" (UID: \"937dc9ee-a952-4496-9659-b4cc9299018a\") " pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.266986 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrgmf\" (UniqueName: \"kubernetes.io/projected/937dc9ee-a952-4496-9659-b4cc9299018a-kube-api-access-qrgmf\") pod \"neutron-d3cd-account-create-update-4ccdt\" (UID: \"937dc9ee-a952-4496-9659-b4cc9299018a\") " pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.267070 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p642p\" (UniqueName: \"kubernetes.io/projected/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-kube-api-access-p642p\") pod \"neutron-db-create-z57pq\" (UID: \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\") " pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.268240 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-operator-scripts\") pod \"neutron-db-create-z57pq\" (UID: \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\") " pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.301535 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p642p\" (UniqueName: \"kubernetes.io/projected/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-kube-api-access-p642p\") pod \"neutron-db-create-z57pq\" (UID: \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\") " pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.368149 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.368997 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937dc9ee-a952-4496-9659-b4cc9299018a-operator-scripts\") pod \"neutron-d3cd-account-create-update-4ccdt\" (UID: \"937dc9ee-a952-4496-9659-b4cc9299018a\") " pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.369055 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrgmf\" (UniqueName: \"kubernetes.io/projected/937dc9ee-a952-4496-9659-b4cc9299018a-kube-api-access-qrgmf\") pod \"neutron-d3cd-account-create-update-4ccdt\" (UID: \"937dc9ee-a952-4496-9659-b4cc9299018a\") " pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.370026 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937dc9ee-a952-4496-9659-b4cc9299018a-operator-scripts\") pod \"neutron-d3cd-account-create-update-4ccdt\" (UID: \"937dc9ee-a952-4496-9659-b4cc9299018a\") " pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.379332 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-jpx62"] Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.390446 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrgmf\" (UniqueName: \"kubernetes.io/projected/937dc9ee-a952-4496-9659-b4cc9299018a-kube-api-access-qrgmf\") pod \"neutron-d3cd-account-create-update-4ccdt\" (UID: \"937dc9ee-a952-4496-9659-b4cc9299018a\") " pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.468333 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.636611 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-fp4tg"] Dec 17 08:38:42 crc kubenswrapper[4966]: I1217 08:38:42.902844 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.000518 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-db-sync-config-data\") pod \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.000566 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khtfj\" (UniqueName: \"kubernetes.io/projected/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-kube-api-access-khtfj\") pod \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.000686 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-combined-ca-bundle\") pod \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.000792 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-config-data\") pod \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\" (UID: \"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06\") " Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.009029 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-e584-account-create-update-hk75s"] Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.011086 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-kube-api-access-khtfj" (OuterVolumeSpecName: "kube-api-access-khtfj") pod "99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" (UID: "99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06"). InnerVolumeSpecName "kube-api-access-khtfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.011087 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" (UID: "99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.025050 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f803-account-create-update-p88mq"] Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.046090 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vp2ln" event={"ID":"99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06","Type":"ContainerDied","Data":"4f059ee8e957ed0e3c118ec50c204493396b71615b51d5b5039a6158fbfba5d4"} Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.046158 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f059ee8e957ed0e3c118ec50c204493396b71615b51d5b5039a6158fbfba5d4" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.046259 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vp2ln" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.054741 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fp4tg" event={"ID":"a952c157-89d5-4bb4-a624-b5b43babd9e0","Type":"ContainerStarted","Data":"c8871e44c031a46565db412b4f6763a3f5d2a2f81ff2916265691c064e982451"} Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.063926 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-jpx62" event={"ID":"23a69382-8e08-47cf-a9aa-bc254229aa21","Type":"ContainerStarted","Data":"1808d23a4649f555ebc6d720fd7f0d7a47c89d9322352250bdbf4bef4a49fa32"} Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.063965 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-jpx62" event={"ID":"23a69382-8e08-47cf-a9aa-bc254229aa21","Type":"ContainerStarted","Data":"a86e3f49d1cb82c74638ee61b27c78804b40f0bd49a6808e82a9520ad36c4457"} Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.097630 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-fp4tg" podStartSLOduration=2.097611968 podStartE2EDuration="2.097611968s" podCreationTimestamp="2025-12-17 08:38:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:43.069310972 +0000 UTC m=+1058.614380904" watchObservedRunningTime="2025-12-17 08:38:43.097611968 +0000 UTC m=+1058.642681920" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.101688 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-jpx62" podStartSLOduration=2.10167508 podStartE2EDuration="2.10167508s" podCreationTimestamp="2025-12-17 08:38:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:43.095290774 +0000 UTC m=+1058.640360716" watchObservedRunningTime="2025-12-17 08:38:43.10167508 +0000 UTC m=+1058.646745012" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.104181 4966 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.104207 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khtfj\" (UniqueName: \"kubernetes.io/projected/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-kube-api-access-khtfj\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.107311 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" (UID: "99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.121654 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-config-data" (OuterVolumeSpecName: "config-data") pod "99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" (UID: "99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.215640 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.215667 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.290401 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d3cd-account-create-update-4ccdt"] Dec 17 08:38:43 crc kubenswrapper[4966]: W1217 08:38:43.416077 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8603b503_2df3_4a05_84e7_024e63d2b455.slice/crio-93aa999ee38a0755166359a94c9d063fa6c400aeed1087fc9469e8b3d32ab7b3 WatchSource:0}: Error finding container 93aa999ee38a0755166359a94c9d063fa6c400aeed1087fc9469e8b3d32ab7b3: Status 404 returned error can't find the container with id 93aa999ee38a0755166359a94c9d063fa6c400aeed1087fc9469e8b3d32ab7b3 Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.444319 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-s8bmz"] Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.504446 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-952b-account-create-update-p5444"] Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.540735 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rjxqq"] Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.568946 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-z57pq"] Dec 17 08:38:43 crc kubenswrapper[4966]: W1217 08:38:43.678972 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f4b0470_0072_4fdc_8a49_33b8c5827ba0.slice/crio-492406dc6b7a0afb25e683461771b2eab4d6a643b22ed01b35718df5ce6014a1 WatchSource:0}: Error finding container 492406dc6b7a0afb25e683461771b2eab4d6a643b22ed01b35718df5ce6014a1: Status 404 returned error can't find the container with id 492406dc6b7a0afb25e683461771b2eab4d6a643b22ed01b35718df5ce6014a1 Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.691783 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58d6b88575-cwxzl"] Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.692025 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" podUID="8c41e1fa-cfd8-466e-b775-d500b0cc4f54" containerName="dnsmasq-dns" containerID="cri-o://75e21cd23cff7c9ef9f9ce1502b45476e56ea6e1655c148b2917ff3c8014ad85" gracePeriod=10 Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.718023 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.779007 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b75d776ff-f25pq"] Dec 17 08:38:43 crc kubenswrapper[4966]: E1217 08:38:43.779352 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" containerName="glance-db-sync" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.779362 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" containerName="glance-db-sync" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.779540 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" containerName="glance-db-sync" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.780357 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.865107 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b75d776ff-f25pq"] Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.889099 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw457\" (UniqueName: \"kubernetes.io/projected/28af8586-3f33-42a1-813b-a1382cde5ec9-kube-api-access-kw457\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.889271 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-config\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.889329 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-svc\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.889383 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-nb\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.889410 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-sb\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.889427 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-swift-storage-0\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.990475 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw457\" (UniqueName: \"kubernetes.io/projected/28af8586-3f33-42a1-813b-a1382cde5ec9-kube-api-access-kw457\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.990549 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-config\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.990579 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-svc\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.990618 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-nb\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.990633 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-sb\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.990650 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-swift-storage-0\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.991678 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-swift-storage-0\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.992704 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-nb\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.993272 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-svc\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.994846 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-sb\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:43 crc kubenswrapper[4966]: I1217 08:38:43.994864 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-config\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.046291 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw457\" (UniqueName: \"kubernetes.io/projected/28af8586-3f33-42a1-813b-a1382cde5ec9-kube-api-access-kw457\") pod \"dnsmasq-dns-b75d776ff-f25pq\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.092830 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fp4tg" event={"ID":"a952c157-89d5-4bb4-a624-b5b43babd9e0","Type":"ContainerStarted","Data":"f79229927e2f10f0d9db670083d2f547170bbb349303724473992f2a2dfec5de"} Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.097919 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s8bmz" event={"ID":"8603b503-2df3-4a05-84e7-024e63d2b455","Type":"ContainerStarted","Data":"93aa999ee38a0755166359a94c9d063fa6c400aeed1087fc9469e8b3d32ab7b3"} Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.101319 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z57pq" event={"ID":"4f4b0470-0072-4fdc-8a49-33b8c5827ba0","Type":"ContainerStarted","Data":"492406dc6b7a0afb25e683461771b2eab4d6a643b22ed01b35718df5ce6014a1"} Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.111910 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f803-account-create-update-p88mq" event={"ID":"552f3591-2913-4bbb-bf5b-79ea4ba40778","Type":"ContainerStarted","Data":"df7136a2732cb708dadd846087872c24276e5b9bd9170e910f84233dbacc4b34"} Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.129904 4966 generic.go:334] "Generic (PLEG): container finished" podID="8c41e1fa-cfd8-466e-b775-d500b0cc4f54" containerID="75e21cd23cff7c9ef9f9ce1502b45476e56ea6e1655c148b2917ff3c8014ad85" exitCode=0 Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.130010 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" event={"ID":"8c41e1fa-cfd8-466e-b775-d500b0cc4f54","Type":"ContainerDied","Data":"75e21cd23cff7c9ef9f9ce1502b45476e56ea6e1655c148b2917ff3c8014ad85"} Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.151037 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-952b-account-create-update-p5444" event={"ID":"771eea8c-0e4e-471c-be4b-3f18e3d91adb","Type":"ContainerStarted","Data":"97d897cc8e2ce0d7b37cf6b0920b504788e6e8a18d23cf7f16453024877e4503"} Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.152620 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rjxqq" event={"ID":"be459859-4e2c-4548-a544-39bf15a731b0","Type":"ContainerStarted","Data":"7d296ff7772aab7b26cf048f1d5a6846d7578b887c53c71fa3f238051c016590"} Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.154325 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d3cd-account-create-update-4ccdt" event={"ID":"937dc9ee-a952-4496-9659-b4cc9299018a","Type":"ContainerStarted","Data":"921eb18315bcfd71e52e94b364b77cf622a141b958cd97913883e859ac0c09db"} Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.156208 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-e584-account-create-update-hk75s" event={"ID":"f70fc4b4-6cf8-4852-827a-d2939c1b656a","Type":"ContainerStarted","Data":"9d041c358e483cc10d0800f52993c0592ef666f0499074149f43736303ac1d87"} Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.191642 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-e584-account-create-update-hk75s" podStartSLOduration=3.191616064 podStartE2EDuration="3.191616064s" podCreationTimestamp="2025-12-17 08:38:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:44.176632333 +0000 UTC m=+1059.721702295" watchObservedRunningTime="2025-12-17 08:38:44.191616064 +0000 UTC m=+1059.736686036" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.409925 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.610776 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.705238 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-sb\") pod \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.705335 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-swift-storage-0\") pod \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.705360 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f7pc\" (UniqueName: \"kubernetes.io/projected/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-kube-api-access-5f7pc\") pod \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.705402 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-nb\") pod \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.705448 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-config\") pod \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.705472 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-svc\") pod \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\" (UID: \"8c41e1fa-cfd8-466e-b775-d500b0cc4f54\") " Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.720563 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-kube-api-access-5f7pc" (OuterVolumeSpecName: "kube-api-access-5f7pc") pod "8c41e1fa-cfd8-466e-b775-d500b0cc4f54" (UID: "8c41e1fa-cfd8-466e-b775-d500b0cc4f54"). InnerVolumeSpecName "kube-api-access-5f7pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.809141 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f7pc\" (UniqueName: \"kubernetes.io/projected/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-kube-api-access-5f7pc\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.823059 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8c41e1fa-cfd8-466e-b775-d500b0cc4f54" (UID: "8c41e1fa-cfd8-466e-b775-d500b0cc4f54"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.827104 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8c41e1fa-cfd8-466e-b775-d500b0cc4f54" (UID: "8c41e1fa-cfd8-466e-b775-d500b0cc4f54"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.851889 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8c41e1fa-cfd8-466e-b775-d500b0cc4f54" (UID: "8c41e1fa-cfd8-466e-b775-d500b0cc4f54"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.868063 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8c41e1fa-cfd8-466e-b775-d500b0cc4f54" (UID: "8c41e1fa-cfd8-466e-b775-d500b0cc4f54"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.883715 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-config" (OuterVolumeSpecName: "config") pod "8c41e1fa-cfd8-466e-b775-d500b0cc4f54" (UID: "8c41e1fa-cfd8-466e-b775-d500b0cc4f54"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.910337 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.910365 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.910375 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.910386 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.910401 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c41e1fa-cfd8-466e-b775-d500b0cc4f54-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:44 crc kubenswrapper[4966]: I1217 08:38:44.986108 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b75d776ff-f25pq"] Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.167024 4966 generic.go:334] "Generic (PLEG): container finished" podID="f70fc4b4-6cf8-4852-827a-d2939c1b656a" containerID="4eb0ce01ae7d7c0e690b982e63d34832ef0d648791fa202b9c8dfbcad882c051" exitCode=0 Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.167099 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-e584-account-create-update-hk75s" event={"ID":"f70fc4b4-6cf8-4852-827a-d2939c1b656a","Type":"ContainerDied","Data":"4eb0ce01ae7d7c0e690b982e63d34832ef0d648791fa202b9c8dfbcad882c051"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.171970 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" event={"ID":"8c41e1fa-cfd8-466e-b775-d500b0cc4f54","Type":"ContainerDied","Data":"431ca7587cf62aa75ca02ce1a569cb1667f49a3391d891bc590cf1212e881ddd"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.172010 4966 scope.go:117] "RemoveContainer" containerID="75e21cd23cff7c9ef9f9ce1502b45476e56ea6e1655c148b2917ff3c8014ad85" Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.172115 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d6b88575-cwxzl" Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.177661 4966 generic.go:334] "Generic (PLEG): container finished" podID="a952c157-89d5-4bb4-a624-b5b43babd9e0" containerID="f79229927e2f10f0d9db670083d2f547170bbb349303724473992f2a2dfec5de" exitCode=0 Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.177743 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fp4tg" event={"ID":"a952c157-89d5-4bb4-a624-b5b43babd9e0","Type":"ContainerDied","Data":"f79229927e2f10f0d9db670083d2f547170bbb349303724473992f2a2dfec5de"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.183689 4966 generic.go:334] "Generic (PLEG): container finished" podID="771eea8c-0e4e-471c-be4b-3f18e3d91adb" containerID="9c2e22d6898f96ac40e0ae9f555d1a0fd0ce5e52652f637dd40568d743b52828" exitCode=0 Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.183746 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-952b-account-create-update-p5444" event={"ID":"771eea8c-0e4e-471c-be4b-3f18e3d91adb","Type":"ContainerDied","Data":"9c2e22d6898f96ac40e0ae9f555d1a0fd0ce5e52652f637dd40568d743b52828"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.186834 4966 generic.go:334] "Generic (PLEG): container finished" podID="be459859-4e2c-4548-a544-39bf15a731b0" containerID="bc37c940d29882f401364b24328941b8b0b13f877bc7f4674dd8a643e6fd34a4" exitCode=0 Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.186896 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rjxqq" event={"ID":"be459859-4e2c-4548-a544-39bf15a731b0","Type":"ContainerDied","Data":"bc37c940d29882f401364b24328941b8b0b13f877bc7f4674dd8a643e6fd34a4"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.189879 4966 generic.go:334] "Generic (PLEG): container finished" podID="23a69382-8e08-47cf-a9aa-bc254229aa21" containerID="1808d23a4649f555ebc6d720fd7f0d7a47c89d9322352250bdbf4bef4a49fa32" exitCode=0 Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.189894 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-jpx62" event={"ID":"23a69382-8e08-47cf-a9aa-bc254229aa21","Type":"ContainerDied","Data":"1808d23a4649f555ebc6d720fd7f0d7a47c89d9322352250bdbf4bef4a49fa32"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.191947 4966 generic.go:334] "Generic (PLEG): container finished" podID="937dc9ee-a952-4496-9659-b4cc9299018a" containerID="44520bfad7d211b8f12aaaf0e2ff9e26735d211290d564e51b91d0c9641fe6fe" exitCode=0 Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.192063 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d3cd-account-create-update-4ccdt" event={"ID":"937dc9ee-a952-4496-9659-b4cc9299018a","Type":"ContainerDied","Data":"44520bfad7d211b8f12aaaf0e2ff9e26735d211290d564e51b91d0c9641fe6fe"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.193503 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" event={"ID":"28af8586-3f33-42a1-813b-a1382cde5ec9","Type":"ContainerStarted","Data":"a01a4e8019ab88346ae2f90c36181219f46649c45a8b98cae5c1eadc7365bcec"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.196621 4966 generic.go:334] "Generic (PLEG): container finished" podID="4f4b0470-0072-4fdc-8a49-33b8c5827ba0" containerID="8adae0b53e1d6b72b2493930742ba86d80089e13ff59ca29f7d345e4edc0c72f" exitCode=0 Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.196666 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z57pq" event={"ID":"4f4b0470-0072-4fdc-8a49-33b8c5827ba0","Type":"ContainerDied","Data":"8adae0b53e1d6b72b2493930742ba86d80089e13ff59ca29f7d345e4edc0c72f"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.204164 4966 generic.go:334] "Generic (PLEG): container finished" podID="552f3591-2913-4bbb-bf5b-79ea4ba40778" containerID="bd94580d0d11729b13ffc6442cf457c306f3630648f4ea98bfa3a1e4d4d84037" exitCode=0 Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.204214 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f803-account-create-update-p88mq" event={"ID":"552f3591-2913-4bbb-bf5b-79ea4ba40778","Type":"ContainerDied","Data":"bd94580d0d11729b13ffc6442cf457c306f3630648f4ea98bfa3a1e4d4d84037"} Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.218217 4966 scope.go:117] "RemoveContainer" containerID="7c97bb4b153934cc04531a97e03bc2a3177f9b31f75a0e8e6f7e978e4dbcf97f" Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.304699 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58d6b88575-cwxzl"] Dec 17 08:38:45 crc kubenswrapper[4966]: I1217 08:38:45.317094 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58d6b88575-cwxzl"] Dec 17 08:38:46 crc kubenswrapper[4966]: I1217 08:38:46.217299 4966 generic.go:334] "Generic (PLEG): container finished" podID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerID="c135e436514d63ea87d326e287784898342c3f76fa4248d93984bc19d0699d82" exitCode=0 Dec 17 08:38:46 crc kubenswrapper[4966]: I1217 08:38:46.217398 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" event={"ID":"28af8586-3f33-42a1-813b-a1382cde5ec9","Type":"ContainerDied","Data":"c135e436514d63ea87d326e287784898342c3f76fa4248d93984bc19d0699d82"} Dec 17 08:38:46 crc kubenswrapper[4966]: I1217 08:38:46.814013 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:38:46 crc kubenswrapper[4966]: I1217 08:38:46.814490 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:38:46 crc kubenswrapper[4966]: I1217 08:38:46.814534 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:38:46 crc kubenswrapper[4966]: I1217 08:38:46.815181 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfe84a49239318b4c24aec015fc04c7e7ced77b13085152134e648f1545c4edf"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 08:38:46 crc kubenswrapper[4966]: I1217 08:38:46.815224 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://dfe84a49239318b4c24aec015fc04c7e7ced77b13085152134e648f1545c4edf" gracePeriod=600 Dec 17 08:38:46 crc kubenswrapper[4966]: I1217 08:38:46.858079 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c41e1fa-cfd8-466e-b775-d500b0cc4f54" path="/var/lib/kubelet/pods/8c41e1fa-cfd8-466e-b775-d500b0cc4f54/volumes" Dec 17 08:38:47 crc kubenswrapper[4966]: I1217 08:38:47.230975 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="dfe84a49239318b4c24aec015fc04c7e7ced77b13085152134e648f1545c4edf" exitCode=0 Dec 17 08:38:47 crc kubenswrapper[4966]: I1217 08:38:47.231065 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"dfe84a49239318b4c24aec015fc04c7e7ced77b13085152134e648f1545c4edf"} Dec 17 08:38:47 crc kubenswrapper[4966]: I1217 08:38:47.231138 4966 scope.go:117] "RemoveContainer" containerID="6002081a3e53c22c985a84be589971720349cfd91c47bfefcfa37db148bb69f7" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.223960 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.255125 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.262194 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.264419 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f803-account-create-update-p88mq" event={"ID":"552f3591-2913-4bbb-bf5b-79ea4ba40778","Type":"ContainerDied","Data":"df7136a2732cb708dadd846087872c24276e5b9bd9170e910f84233dbacc4b34"} Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.264445 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df7136a2732cb708dadd846087872c24276e5b9bd9170e910f84233dbacc4b34" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.265755 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-952b-account-create-update-p5444" event={"ID":"771eea8c-0e4e-471c-be4b-3f18e3d91adb","Type":"ContainerDied","Data":"97d897cc8e2ce0d7b37cf6b0920b504788e6e8a18d23cf7f16453024877e4503"} Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.265777 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97d897cc8e2ce0d7b37cf6b0920b504788e6e8a18d23cf7f16453024877e4503" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.269638 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rjxqq" event={"ID":"be459859-4e2c-4548-a544-39bf15a731b0","Type":"ContainerDied","Data":"7d296ff7772aab7b26cf048f1d5a6846d7578b887c53c71fa3f238051c016590"} Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.269674 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d296ff7772aab7b26cf048f1d5a6846d7578b887c53c71fa3f238051c016590" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.271112 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.291390 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-jpx62" event={"ID":"23a69382-8e08-47cf-a9aa-bc254229aa21","Type":"ContainerDied","Data":"a86e3f49d1cb82c74638ee61b27c78804b40f0bd49a6808e82a9520ad36c4457"} Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.291430 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a86e3f49d1cb82c74638ee61b27c78804b40f0bd49a6808e82a9520ad36c4457" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.293249 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d3cd-account-create-update-4ccdt" event={"ID":"937dc9ee-a952-4496-9659-b4cc9299018a","Type":"ContainerDied","Data":"921eb18315bcfd71e52e94b364b77cf622a141b958cd97913883e859ac0c09db"} Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.293269 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="921eb18315bcfd71e52e94b364b77cf622a141b958cd97913883e859ac0c09db" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.293322 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d3cd-account-create-update-4ccdt" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.302246 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e584-account-create-update-hk75s" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.302257 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-e584-account-create-update-hk75s" event={"ID":"f70fc4b4-6cf8-4852-827a-d2939c1b656a","Type":"ContainerDied","Data":"9d041c358e483cc10d0800f52993c0592ef666f0499074149f43736303ac1d87"} Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.302581 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d041c358e483cc10d0800f52993c0592ef666f0499074149f43736303ac1d87" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.303520 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fp4tg" event={"ID":"a952c157-89d5-4bb4-a624-b5b43babd9e0","Type":"ContainerDied","Data":"c8871e44c031a46565db412b4f6763a3f5d2a2f81ff2916265691c064e982451"} Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.303535 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8871e44c031a46565db412b4f6763a3f5d2a2f81ff2916265691c064e982451" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.303588 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fp4tg" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.315442 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.330136 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.338014 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z57pq" event={"ID":"4f4b0470-0072-4fdc-8a49-33b8c5827ba0","Type":"ContainerDied","Data":"492406dc6b7a0afb25e683461771b2eab4d6a643b22ed01b35718df5ce6014a1"} Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.338063 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="492406dc6b7a0afb25e683461771b2eab4d6a643b22ed01b35718df5ce6014a1" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.338131 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z57pq" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.363422 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.363589 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-jpx62" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.401841 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9jj6\" (UniqueName: \"kubernetes.io/projected/be459859-4e2c-4548-a544-39bf15a731b0-kube-api-access-w9jj6\") pod \"be459859-4e2c-4548-a544-39bf15a731b0\" (UID: \"be459859-4e2c-4548-a544-39bf15a731b0\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.401911 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be459859-4e2c-4548-a544-39bf15a731b0-operator-scripts\") pod \"be459859-4e2c-4548-a544-39bf15a731b0\" (UID: \"be459859-4e2c-4548-a544-39bf15a731b0\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.401931 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spxqp\" (UniqueName: \"kubernetes.io/projected/f70fc4b4-6cf8-4852-827a-d2939c1b656a-kube-api-access-spxqp\") pod \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\" (UID: \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.401988 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck6kz\" (UniqueName: \"kubernetes.io/projected/a952c157-89d5-4bb4-a624-b5b43babd9e0-kube-api-access-ck6kz\") pod \"a952c157-89d5-4bb4-a624-b5b43babd9e0\" (UID: \"a952c157-89d5-4bb4-a624-b5b43babd9e0\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402028 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937dc9ee-a952-4496-9659-b4cc9299018a-operator-scripts\") pod \"937dc9ee-a952-4496-9659-b4cc9299018a\" (UID: \"937dc9ee-a952-4496-9659-b4cc9299018a\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402072 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv7sg\" (UniqueName: \"kubernetes.io/projected/552f3591-2913-4bbb-bf5b-79ea4ba40778-kube-api-access-rv7sg\") pod \"552f3591-2913-4bbb-bf5b-79ea4ba40778\" (UID: \"552f3591-2913-4bbb-bf5b-79ea4ba40778\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402104 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23a69382-8e08-47cf-a9aa-bc254229aa21-operator-scripts\") pod \"23a69382-8e08-47cf-a9aa-bc254229aa21\" (UID: \"23a69382-8e08-47cf-a9aa-bc254229aa21\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402150 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrgmf\" (UniqueName: \"kubernetes.io/projected/937dc9ee-a952-4496-9659-b4cc9299018a-kube-api-access-qrgmf\") pod \"937dc9ee-a952-4496-9659-b4cc9299018a\" (UID: \"937dc9ee-a952-4496-9659-b4cc9299018a\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402181 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bb8x\" (UniqueName: \"kubernetes.io/projected/771eea8c-0e4e-471c-be4b-3f18e3d91adb-kube-api-access-6bb8x\") pod \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\" (UID: \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402223 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p642p\" (UniqueName: \"kubernetes.io/projected/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-kube-api-access-p642p\") pod \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\" (UID: \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402301 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a952c157-89d5-4bb4-a624-b5b43babd9e0-operator-scripts\") pod \"a952c157-89d5-4bb4-a624-b5b43babd9e0\" (UID: \"a952c157-89d5-4bb4-a624-b5b43babd9e0\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402338 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/771eea8c-0e4e-471c-be4b-3f18e3d91adb-operator-scripts\") pod \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\" (UID: \"771eea8c-0e4e-471c-be4b-3f18e3d91adb\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402379 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552f3591-2913-4bbb-bf5b-79ea4ba40778-operator-scripts\") pod \"552f3591-2913-4bbb-bf5b-79ea4ba40778\" (UID: \"552f3591-2913-4bbb-bf5b-79ea4ba40778\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402490 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70fc4b4-6cf8-4852-827a-d2939c1b656a-operator-scripts\") pod \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\" (UID: \"f70fc4b4-6cf8-4852-827a-d2939c1b656a\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402611 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwgvt\" (UniqueName: \"kubernetes.io/projected/23a69382-8e08-47cf-a9aa-bc254229aa21-kube-api-access-dwgvt\") pod \"23a69382-8e08-47cf-a9aa-bc254229aa21\" (UID: \"23a69382-8e08-47cf-a9aa-bc254229aa21\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.402646 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-operator-scripts\") pod \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\" (UID: \"4f4b0470-0072-4fdc-8a49-33b8c5827ba0\") " Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.403936 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4f4b0470-0072-4fdc-8a49-33b8c5827ba0" (UID: "4f4b0470-0072-4fdc-8a49-33b8c5827ba0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.404863 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be459859-4e2c-4548-a544-39bf15a731b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "be459859-4e2c-4548-a544-39bf15a731b0" (UID: "be459859-4e2c-4548-a544-39bf15a731b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.410147 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/937dc9ee-a952-4496-9659-b4cc9299018a-kube-api-access-qrgmf" (OuterVolumeSpecName: "kube-api-access-qrgmf") pod "937dc9ee-a952-4496-9659-b4cc9299018a" (UID: "937dc9ee-a952-4496-9659-b4cc9299018a"). InnerVolumeSpecName "kube-api-access-qrgmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.410613 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/771eea8c-0e4e-471c-be4b-3f18e3d91adb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "771eea8c-0e4e-471c-be4b-3f18e3d91adb" (UID: "771eea8c-0e4e-471c-be4b-3f18e3d91adb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.410658 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f70fc4b4-6cf8-4852-827a-d2939c1b656a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f70fc4b4-6cf8-4852-827a-d2939c1b656a" (UID: "f70fc4b4-6cf8-4852-827a-d2939c1b656a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.410660 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a952c157-89d5-4bb4-a624-b5b43babd9e0-kube-api-access-ck6kz" (OuterVolumeSpecName: "kube-api-access-ck6kz") pod "a952c157-89d5-4bb4-a624-b5b43babd9e0" (UID: "a952c157-89d5-4bb4-a624-b5b43babd9e0"). InnerVolumeSpecName "kube-api-access-ck6kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.411068 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be459859-4e2c-4548-a544-39bf15a731b0-kube-api-access-w9jj6" (OuterVolumeSpecName: "kube-api-access-w9jj6") pod "be459859-4e2c-4548-a544-39bf15a731b0" (UID: "be459859-4e2c-4548-a544-39bf15a731b0"). InnerVolumeSpecName "kube-api-access-w9jj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.411123 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/937dc9ee-a952-4496-9659-b4cc9299018a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "937dc9ee-a952-4496-9659-b4cc9299018a" (UID: "937dc9ee-a952-4496-9659-b4cc9299018a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.411175 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23a69382-8e08-47cf-a9aa-bc254229aa21-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "23a69382-8e08-47cf-a9aa-bc254229aa21" (UID: "23a69382-8e08-47cf-a9aa-bc254229aa21"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.411486 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a952c157-89d5-4bb4-a624-b5b43babd9e0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a952c157-89d5-4bb4-a624-b5b43babd9e0" (UID: "a952c157-89d5-4bb4-a624-b5b43babd9e0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.411551 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/552f3591-2913-4bbb-bf5b-79ea4ba40778-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "552f3591-2913-4bbb-bf5b-79ea4ba40778" (UID: "552f3591-2913-4bbb-bf5b-79ea4ba40778"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.422042 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23a69382-8e08-47cf-a9aa-bc254229aa21-kube-api-access-dwgvt" (OuterVolumeSpecName: "kube-api-access-dwgvt") pod "23a69382-8e08-47cf-a9aa-bc254229aa21" (UID: "23a69382-8e08-47cf-a9aa-bc254229aa21"). InnerVolumeSpecName "kube-api-access-dwgvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.426439 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-kube-api-access-p642p" (OuterVolumeSpecName: "kube-api-access-p642p") pod "4f4b0470-0072-4fdc-8a49-33b8c5827ba0" (UID: "4f4b0470-0072-4fdc-8a49-33b8c5827ba0"). InnerVolumeSpecName "kube-api-access-p642p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.427734 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f70fc4b4-6cf8-4852-827a-d2939c1b656a-kube-api-access-spxqp" (OuterVolumeSpecName: "kube-api-access-spxqp") pod "f70fc4b4-6cf8-4852-827a-d2939c1b656a" (UID: "f70fc4b4-6cf8-4852-827a-d2939c1b656a"). InnerVolumeSpecName "kube-api-access-spxqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.427854 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/771eea8c-0e4e-471c-be4b-3f18e3d91adb-kube-api-access-6bb8x" (OuterVolumeSpecName: "kube-api-access-6bb8x") pod "771eea8c-0e4e-471c-be4b-3f18e3d91adb" (UID: "771eea8c-0e4e-471c-be4b-3f18e3d91adb"). InnerVolumeSpecName "kube-api-access-6bb8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.430726 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/552f3591-2913-4bbb-bf5b-79ea4ba40778-kube-api-access-rv7sg" (OuterVolumeSpecName: "kube-api-access-rv7sg") pod "552f3591-2913-4bbb-bf5b-79ea4ba40778" (UID: "552f3591-2913-4bbb-bf5b-79ea4ba40778"). InnerVolumeSpecName "kube-api-access-rv7sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504528 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a952c157-89d5-4bb4-a624-b5b43babd9e0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504580 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/771eea8c-0e4e-471c-be4b-3f18e3d91adb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504593 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552f3591-2913-4bbb-bf5b-79ea4ba40778-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504603 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70fc4b4-6cf8-4852-827a-d2939c1b656a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504614 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwgvt\" (UniqueName: \"kubernetes.io/projected/23a69382-8e08-47cf-a9aa-bc254229aa21-kube-api-access-dwgvt\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504626 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504637 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9jj6\" (UniqueName: \"kubernetes.io/projected/be459859-4e2c-4548-a544-39bf15a731b0-kube-api-access-w9jj6\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504646 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be459859-4e2c-4548-a544-39bf15a731b0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504658 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spxqp\" (UniqueName: \"kubernetes.io/projected/f70fc4b4-6cf8-4852-827a-d2939c1b656a-kube-api-access-spxqp\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504668 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck6kz\" (UniqueName: \"kubernetes.io/projected/a952c157-89d5-4bb4-a624-b5b43babd9e0-kube-api-access-ck6kz\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504677 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/937dc9ee-a952-4496-9659-b4cc9299018a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504687 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv7sg\" (UniqueName: \"kubernetes.io/projected/552f3591-2913-4bbb-bf5b-79ea4ba40778-kube-api-access-rv7sg\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504697 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23a69382-8e08-47cf-a9aa-bc254229aa21-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504707 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrgmf\" (UniqueName: \"kubernetes.io/projected/937dc9ee-a952-4496-9659-b4cc9299018a-kube-api-access-qrgmf\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504717 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bb8x\" (UniqueName: \"kubernetes.io/projected/771eea8c-0e4e-471c-be4b-3f18e3d91adb-kube-api-access-6bb8x\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:49 crc kubenswrapper[4966]: I1217 08:38:49.504728 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p642p\" (UniqueName: \"kubernetes.io/projected/4f4b0470-0072-4fdc-8a49-33b8c5827ba0-kube-api-access-p642p\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.353403 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" event={"ID":"28af8586-3f33-42a1-813b-a1382cde5ec9","Type":"ContainerStarted","Data":"f54a4f58f7bb67f9d87e7addb7489a5b71d139a3f9c89fc2c158fe0bbd974742"} Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.353949 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.357917 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s8bmz" event={"ID":"8603b503-2df3-4a05-84e7-024e63d2b455","Type":"ContainerStarted","Data":"79f04cd102f2163311766ff25fb8b27cb5227115c3882b44f13422cee82b6b26"} Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.363335 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-jpx62" Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.372274 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"ede17bc7a2bbfa2ac454efc1ee7643bc3db24c460d6b848c803354fff3d0b2f2"} Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.372916 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f803-account-create-update-p88mq" Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.372985 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-952b-account-create-update-p5444" Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.373349 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rjxqq" Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.416281 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" podStartSLOduration=7.41625522 podStartE2EDuration="7.41625522s" podCreationTimestamp="2025-12-17 08:38:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:50.387678506 +0000 UTC m=+1065.932748458" watchObservedRunningTime="2025-12-17 08:38:50.41625522 +0000 UTC m=+1065.961325162" Dec 17 08:38:50 crc kubenswrapper[4966]: I1217 08:38:50.434916 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-s8bmz" podStartSLOduration=3.791406705 podStartE2EDuration="9.434898181s" podCreationTimestamp="2025-12-17 08:38:41 +0000 UTC" firstStartedPulling="2025-12-17 08:38:43.420990952 +0000 UTC m=+1058.966060894" lastFinishedPulling="2025-12-17 08:38:49.064482408 +0000 UTC m=+1064.609552370" observedRunningTime="2025-12-17 08:38:50.413903405 +0000 UTC m=+1065.958973347" watchObservedRunningTime="2025-12-17 08:38:50.434898181 +0000 UTC m=+1065.979968123" Dec 17 08:38:54 crc kubenswrapper[4966]: I1217 08:38:54.410501 4966 generic.go:334] "Generic (PLEG): container finished" podID="8603b503-2df3-4a05-84e7-024e63d2b455" containerID="79f04cd102f2163311766ff25fb8b27cb5227115c3882b44f13422cee82b6b26" exitCode=0 Dec 17 08:38:54 crc kubenswrapper[4966]: I1217 08:38:54.410636 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s8bmz" event={"ID":"8603b503-2df3-4a05-84e7-024e63d2b455","Type":"ContainerDied","Data":"79f04cd102f2163311766ff25fb8b27cb5227115c3882b44f13422cee82b6b26"} Dec 17 08:38:54 crc kubenswrapper[4966]: I1217 08:38:54.412180 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:38:54 crc kubenswrapper[4966]: I1217 08:38:54.522243 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cbbc897ff-7tpcm"] Dec 17 08:38:54 crc kubenswrapper[4966]: I1217 08:38:54.522591 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" podUID="0c2b7867-860e-46c7-9981-55fefdfe7458" containerName="dnsmasq-dns" containerID="cri-o://2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d" gracePeriod=10 Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.010996 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.104678 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsdc7\" (UniqueName: \"kubernetes.io/projected/0c2b7867-860e-46c7-9981-55fefdfe7458-kube-api-access-xsdc7\") pod \"0c2b7867-860e-46c7-9981-55fefdfe7458\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.104767 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-nb\") pod \"0c2b7867-860e-46c7-9981-55fefdfe7458\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.104817 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-sb\") pod \"0c2b7867-860e-46c7-9981-55fefdfe7458\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.104953 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-config\") pod \"0c2b7867-860e-46c7-9981-55fefdfe7458\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.105023 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-dns-svc\") pod \"0c2b7867-860e-46c7-9981-55fefdfe7458\" (UID: \"0c2b7867-860e-46c7-9981-55fefdfe7458\") " Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.109705 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c2b7867-860e-46c7-9981-55fefdfe7458-kube-api-access-xsdc7" (OuterVolumeSpecName: "kube-api-access-xsdc7") pod "0c2b7867-860e-46c7-9981-55fefdfe7458" (UID: "0c2b7867-860e-46c7-9981-55fefdfe7458"). InnerVolumeSpecName "kube-api-access-xsdc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.145682 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c2b7867-860e-46c7-9981-55fefdfe7458" (UID: "0c2b7867-860e-46c7-9981-55fefdfe7458"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.150842 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0c2b7867-860e-46c7-9981-55fefdfe7458" (UID: "0c2b7867-860e-46c7-9981-55fefdfe7458"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.155480 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-config" (OuterVolumeSpecName: "config") pod "0c2b7867-860e-46c7-9981-55fefdfe7458" (UID: "0c2b7867-860e-46c7-9981-55fefdfe7458"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.179031 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0c2b7867-860e-46c7-9981-55fefdfe7458" (UID: "0c2b7867-860e-46c7-9981-55fefdfe7458"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.206939 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.206987 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.207001 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.207014 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsdc7\" (UniqueName: \"kubernetes.io/projected/0c2b7867-860e-46c7-9981-55fefdfe7458-kube-api-access-xsdc7\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.207027 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c2b7867-860e-46c7-9981-55fefdfe7458-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.423146 4966 generic.go:334] "Generic (PLEG): container finished" podID="0c2b7867-860e-46c7-9981-55fefdfe7458" containerID="2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d" exitCode=0 Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.423398 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.425049 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" event={"ID":"0c2b7867-860e-46c7-9981-55fefdfe7458","Type":"ContainerDied","Data":"2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d"} Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.425124 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbbc897ff-7tpcm" event={"ID":"0c2b7867-860e-46c7-9981-55fefdfe7458","Type":"ContainerDied","Data":"508bcc1a6fd61478a53ba81c31a7bc05d3126f5e103c9a42ca61d4077da7d02f"} Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.425148 4966 scope.go:117] "RemoveContainer" containerID="2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.445505 4966 scope.go:117] "RemoveContainer" containerID="e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.472011 4966 scope.go:117] "RemoveContainer" containerID="2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d" Dec 17 08:38:55 crc kubenswrapper[4966]: E1217 08:38:55.473015 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d\": container with ID starting with 2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d not found: ID does not exist" containerID="2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.473070 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d"} err="failed to get container status \"2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d\": rpc error: code = NotFound desc = could not find container \"2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d\": container with ID starting with 2ffea1371076d8f40f09716f1683b4213f54cea6e68e4ba5271791b15fd6d33d not found: ID does not exist" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.473102 4966 scope.go:117] "RemoveContainer" containerID="e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1" Dec 17 08:38:55 crc kubenswrapper[4966]: E1217 08:38:55.473436 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1\": container with ID starting with e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1 not found: ID does not exist" containerID="e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.473564 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1"} err="failed to get container status \"e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1\": rpc error: code = NotFound desc = could not find container \"e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1\": container with ID starting with e4c197f8c1621c3d515655523544abc32d67111b83421f30d245164199c487e1 not found: ID does not exist" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.476707 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cbbc897ff-7tpcm"] Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.486217 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cbbc897ff-7tpcm"] Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.711505 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.818121 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxw99\" (UniqueName: \"kubernetes.io/projected/8603b503-2df3-4a05-84e7-024e63d2b455-kube-api-access-hxw99\") pod \"8603b503-2df3-4a05-84e7-024e63d2b455\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.818305 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-combined-ca-bundle\") pod \"8603b503-2df3-4a05-84e7-024e63d2b455\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.818393 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-config-data\") pod \"8603b503-2df3-4a05-84e7-024e63d2b455\" (UID: \"8603b503-2df3-4a05-84e7-024e63d2b455\") " Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.824418 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8603b503-2df3-4a05-84e7-024e63d2b455-kube-api-access-hxw99" (OuterVolumeSpecName: "kube-api-access-hxw99") pod "8603b503-2df3-4a05-84e7-024e63d2b455" (UID: "8603b503-2df3-4a05-84e7-024e63d2b455"). InnerVolumeSpecName "kube-api-access-hxw99". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.846943 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8603b503-2df3-4a05-84e7-024e63d2b455" (UID: "8603b503-2df3-4a05-84e7-024e63d2b455"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.862081 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-config-data" (OuterVolumeSpecName: "config-data") pod "8603b503-2df3-4a05-84e7-024e63d2b455" (UID: "8603b503-2df3-4a05-84e7-024e63d2b455"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.920738 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.920776 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxw99\" (UniqueName: \"kubernetes.io/projected/8603b503-2df3-4a05-84e7-024e63d2b455-kube-api-access-hxw99\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:55 crc kubenswrapper[4966]: I1217 08:38:55.920787 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8603b503-2df3-4a05-84e7-024e63d2b455-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.432482 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-s8bmz" event={"ID":"8603b503-2df3-4a05-84e7-024e63d2b455","Type":"ContainerDied","Data":"93aa999ee38a0755166359a94c9d063fa6c400aeed1087fc9469e8b3d32ab7b3"} Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.432521 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93aa999ee38a0755166359a94c9d063fa6c400aeed1087fc9469e8b3d32ab7b3" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.432576 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-s8bmz" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713303 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-856cfb64ff-679wx"] Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713641 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70fc4b4-6cf8-4852-827a-d2939c1b656a" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713654 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70fc4b4-6cf8-4852-827a-d2939c1b656a" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713668 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c2b7867-860e-46c7-9981-55fefdfe7458" containerName="dnsmasq-dns" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713674 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c2b7867-860e-46c7-9981-55fefdfe7458" containerName="dnsmasq-dns" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713684 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a952c157-89d5-4bb4-a624-b5b43babd9e0" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713690 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a952c157-89d5-4bb4-a624-b5b43babd9e0" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713706 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="552f3591-2913-4bbb-bf5b-79ea4ba40778" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713711 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="552f3591-2913-4bbb-bf5b-79ea4ba40778" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713719 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be459859-4e2c-4548-a544-39bf15a731b0" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713725 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="be459859-4e2c-4548-a544-39bf15a731b0" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713736 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c41e1fa-cfd8-466e-b775-d500b0cc4f54" containerName="init" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713742 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c41e1fa-cfd8-466e-b775-d500b0cc4f54" containerName="init" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713751 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="937dc9ee-a952-4496-9659-b4cc9299018a" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713758 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="937dc9ee-a952-4496-9659-b4cc9299018a" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713768 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8603b503-2df3-4a05-84e7-024e63d2b455" containerName="keystone-db-sync" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713774 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8603b503-2df3-4a05-84e7-024e63d2b455" containerName="keystone-db-sync" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713789 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c41e1fa-cfd8-466e-b775-d500b0cc4f54" containerName="dnsmasq-dns" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713795 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c41e1fa-cfd8-466e-b775-d500b0cc4f54" containerName="dnsmasq-dns" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713804 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c2b7867-860e-46c7-9981-55fefdfe7458" containerName="init" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713809 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c2b7867-860e-46c7-9981-55fefdfe7458" containerName="init" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713822 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a69382-8e08-47cf-a9aa-bc254229aa21" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713827 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a69382-8e08-47cf-a9aa-bc254229aa21" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713839 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f4b0470-0072-4fdc-8a49-33b8c5827ba0" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713846 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f4b0470-0072-4fdc-8a49-33b8c5827ba0" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: E1217 08:38:56.713855 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="771eea8c-0e4e-471c-be4b-3f18e3d91adb" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.713863 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="771eea8c-0e4e-471c-be4b-3f18e3d91adb" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714013 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f70fc4b4-6cf8-4852-827a-d2939c1b656a" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714036 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a952c157-89d5-4bb4-a624-b5b43babd9e0" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714050 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c2b7867-860e-46c7-9981-55fefdfe7458" containerName="dnsmasq-dns" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714068 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="552f3591-2913-4bbb-bf5b-79ea4ba40778" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714076 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c41e1fa-cfd8-466e-b775-d500b0cc4f54" containerName="dnsmasq-dns" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714082 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="23a69382-8e08-47cf-a9aa-bc254229aa21" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714091 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="8603b503-2df3-4a05-84e7-024e63d2b455" containerName="keystone-db-sync" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714107 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f4b0470-0072-4fdc-8a49-33b8c5827ba0" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714119 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="937dc9ee-a952-4496-9659-b4cc9299018a" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714135 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="771eea8c-0e4e-471c-be4b-3f18e3d91adb" containerName="mariadb-account-create-update" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714144 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="be459859-4e2c-4548-a544-39bf15a731b0" containerName="mariadb-database-create" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.714972 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.738597 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-856cfb64ff-679wx"] Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.768312 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-69m6t"] Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.769906 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.776537 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.776759 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.777008 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.777189 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.777332 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-czvn7" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.784110 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-69m6t"] Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.834981 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-nb\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.835037 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-swift-storage-0\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.835101 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-config\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.835131 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-svc\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.835190 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-sb\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.835273 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ls49\" (UniqueName: \"kubernetes.io/projected/e5a24615-9b89-4a84-b28b-cc5cf0d81682-kube-api-access-2ls49\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.847979 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c2b7867-860e-46c7-9981-55fefdfe7458" path="/var/lib/kubelet/pods/0c2b7867-860e-46c7-9981-55fefdfe7458/volumes" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940041 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-swift-storage-0\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940107 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-config-data\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940164 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-scripts\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940223 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-config\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940271 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-credential-keys\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940295 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-svc\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940326 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-combined-ca-bundle\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940351 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-fernet-keys\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940384 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29mzh\" (UniqueName: \"kubernetes.io/projected/225a53d8-4292-436b-bb51-6527d51180c1-kube-api-access-29mzh\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940430 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-sb\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940469 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ls49\" (UniqueName: \"kubernetes.io/projected/e5a24615-9b89-4a84-b28b-cc5cf0d81682-kube-api-access-2ls49\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940510 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-nb\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.940947 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-swift-storage-0\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.941403 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-nb\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.941779 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-sb\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.941786 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-config\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.942183 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-svc\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.952210 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-9bd96"] Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.953306 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.964277 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-l4hjd" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.964501 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 17 08:38:56 crc kubenswrapper[4966]: I1217 08:38:56.990483 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-9bd96"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:56.999504 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ls49\" (UniqueName: \"kubernetes.io/projected/e5a24615-9b89-4a84-b28b-cc5cf0d81682-kube-api-access-2ls49\") pod \"dnsmasq-dns-856cfb64ff-679wx\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.046058 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-combined-ca-bundle\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.046314 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-fernet-keys\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.046344 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29mzh\" (UniqueName: \"kubernetes.io/projected/225a53d8-4292-436b-bb51-6527d51180c1-kube-api-access-29mzh\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.046393 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-combined-ca-bundle\") pod \"heat-db-sync-9bd96\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.046458 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-config-data\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.046481 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbq6s\" (UniqueName: \"kubernetes.io/projected/b7351028-bc40-4197-baa1-9fb0c525170d-kube-api-access-pbq6s\") pod \"heat-db-sync-9bd96\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.046517 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-config-data\") pod \"heat-db-sync-9bd96\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.046534 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-scripts\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.046563 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-credential-keys\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.065213 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-credential-keys\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.070704 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.080437 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-fernet-keys\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.080651 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-scripts\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.104422 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-combined-ca-bundle\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.130683 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-config-data\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.135477 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29mzh\" (UniqueName: \"kubernetes.io/projected/225a53d8-4292-436b-bb51-6527d51180c1-kube-api-access-29mzh\") pod \"keystone-bootstrap-69m6t\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.136357 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.140932 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-rqvfb"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.142024 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.147598 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-combined-ca-bundle\") pod \"heat-db-sync-9bd96\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.147717 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbq6s\" (UniqueName: \"kubernetes.io/projected/b7351028-bc40-4197-baa1-9fb0c525170d-kube-api-access-pbq6s\") pod \"heat-db-sync-9bd96\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.147760 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-config-data\") pod \"heat-db-sync-9bd96\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.172605 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kg8zx" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.172822 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.172993 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.185727 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-config-data\") pod \"heat-db-sync-9bd96\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.189585 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-combined-ca-bundle\") pod \"heat-db-sync-9bd96\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.191005 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rqvfb"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.205526 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbq6s\" (UniqueName: \"kubernetes.io/projected/b7351028-bc40-4197-baa1-9fb0c525170d-kube-api-access-pbq6s\") pod \"heat-db-sync-9bd96\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.221927 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-fhj4g"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.237639 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.249087 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r6ks\" (UniqueName: \"kubernetes.io/projected/3bca810d-b802-481a-8792-3be9b6a141bc-kube-api-access-9r6ks\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.249178 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-db-sync-config-data\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.249212 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-scripts\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.249250 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-config-data\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.249306 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bca810d-b802-481a-8792-3be9b6a141bc-etc-machine-id\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.249338 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-combined-ca-bundle\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.260732 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.261272 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.262051 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-bx4bj" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.304364 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9bd96" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.308681 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68d6578c4c-8tfk5"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.310206 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.329437 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.329699 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.329851 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-qx7ks" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.330658 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.354095 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-config\") pod \"neutron-db-sync-fhj4g\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.354156 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r6ks\" (UniqueName: \"kubernetes.io/projected/3bca810d-b802-481a-8792-3be9b6a141bc-kube-api-access-9r6ks\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.354193 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-db-sync-config-data\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.354220 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-scripts\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.354246 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlm7l\" (UniqueName: \"kubernetes.io/projected/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-kube-api-access-dlm7l\") pod \"neutron-db-sync-fhj4g\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.354276 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-config-data\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.354483 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-combined-ca-bundle\") pod \"neutron-db-sync-fhj4g\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.354528 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bca810d-b802-481a-8792-3be9b6a141bc-etc-machine-id\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.356183 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-combined-ca-bundle\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.362782 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-config-data\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.364149 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bca810d-b802-481a-8792-3be9b6a141bc-etc-machine-id\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.364291 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fhj4g"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.375157 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-db-sync-config-data\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.377497 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-scripts\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.377640 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-combined-ca-bundle\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.431530 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.433487 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.450368 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.450366 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.469683 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r6ks\" (UniqueName: \"kubernetes.io/projected/3bca810d-b802-481a-8792-3be9b6a141bc-kube-api-access-9r6ks\") pod \"cinder-db-sync-rqvfb\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.469791 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-config\") pod \"neutron-db-sync-fhj4g\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.469834 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-scripts\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.469916 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c007a81-a0c1-43d8-8a77-62aebd2f188d-horizon-secret-key\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.469977 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-config-data\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.470049 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlm7l\" (UniqueName: \"kubernetes.io/projected/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-kube-api-access-dlm7l\") pod \"neutron-db-sync-fhj4g\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.470144 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c007a81-a0c1-43d8-8a77-62aebd2f188d-logs\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.470165 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-combined-ca-bundle\") pod \"neutron-db-sync-fhj4g\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.470186 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqrgz\" (UniqueName: \"kubernetes.io/projected/0c007a81-a0c1-43d8-8a77-62aebd2f188d-kube-api-access-vqrgz\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.476564 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-combined-ca-bundle\") pod \"neutron-db-sync-fhj4g\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.484594 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-config\") pod \"neutron-db-sync-fhj4g\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.558272 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.559155 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlm7l\" (UniqueName: \"kubernetes.io/projected/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-kube-api-access-dlm7l\") pod \"neutron-db-sync-fhj4g\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.574757 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-run-httpd\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.574812 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-log-httpd\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.574835 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.574855 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-config-data\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.574901 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c007a81-a0c1-43d8-8a77-62aebd2f188d-logs\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.574928 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqrgz\" (UniqueName: \"kubernetes.io/projected/0c007a81-a0c1-43d8-8a77-62aebd2f188d-kube-api-access-vqrgz\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.574964 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.575001 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf4xx\" (UniqueName: \"kubernetes.io/projected/b26b94bc-5f74-4321-8e33-28b28b94ac50-kube-api-access-cf4xx\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.575040 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-scripts\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.575071 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c007a81-a0c1-43d8-8a77-62aebd2f188d-horizon-secret-key\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.575099 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-config-data\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.575123 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-scripts\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.575669 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c007a81-a0c1-43d8-8a77-62aebd2f188d-logs\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.584400 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-scripts\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.587794 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-config-data\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.593074 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6jj7z"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.600856 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.602283 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6jj7z"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.610497 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sgmwm" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.610758 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.631805 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c007a81-a0c1-43d8-8a77-62aebd2f188d-horizon-secret-key\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.635232 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.639892 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqrgz\" (UniqueName: \"kubernetes.io/projected/0c007a81-a0c1-43d8-8a77-62aebd2f188d-kube-api-access-vqrgz\") pod \"horizon-68d6578c4c-8tfk5\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.640846 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.684723 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.684807 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp27h\" (UniqueName: \"kubernetes.io/projected/eccb295b-14f1-43a9-98c8-673e2ce78add-kube-api-access-lp27h\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.684834 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf4xx\" (UniqueName: \"kubernetes.io/projected/b26b94bc-5f74-4321-8e33-28b28b94ac50-kube-api-access-cf4xx\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.684886 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-combined-ca-bundle\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.684922 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-scripts\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.684955 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-run-httpd\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.684974 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-scripts\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.684993 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-log-httpd\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.685020 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.685037 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-config-data\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.685053 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-config-data\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.685083 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eccb295b-14f1-43a9-98c8-673e2ce78add-logs\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.687440 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-run-httpd\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.696164 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-log-httpd\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.699232 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9vf55"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.701139 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.706102 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.706331 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.706860 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-c7pcz" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.710598 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-config-data\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.716136 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.717812 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-scripts\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.725777 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.746740 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-856cfb64ff-679wx"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.749030 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.762949 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf4xx\" (UniqueName: \"kubernetes.io/projected/b26b94bc-5f74-4321-8e33-28b28b94ac50-kube-api-access-cf4xx\") pod \"ceilometer-0\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.797534 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-combined-ca-bundle\") pod \"barbican-db-sync-9vf55\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.797581 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-scripts\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.797608 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-config-data\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.797642 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eccb295b-14f1-43a9-98c8-673e2ce78add-logs\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.797669 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-db-sync-config-data\") pod \"barbican-db-sync-9vf55\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.797697 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp27h\" (UniqueName: \"kubernetes.io/projected/eccb295b-14f1-43a9-98c8-673e2ce78add-kube-api-access-lp27h\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.797726 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrr57\" (UniqueName: \"kubernetes.io/projected/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-kube-api-access-xrr57\") pod \"barbican-db-sync-9vf55\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.797755 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-combined-ca-bundle\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.799039 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eccb295b-14f1-43a9-98c8-673e2ce78add-logs\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.814101 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-scripts\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.814416 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-combined-ca-bundle\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.818375 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-config-data\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.818808 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.832284 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp27h\" (UniqueName: \"kubernetes.io/projected/eccb295b-14f1-43a9-98c8-673e2ce78add-kube-api-access-lp27h\") pod \"placement-db-sync-6jj7z\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.844982 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68d6578c4c-8tfk5"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.879685 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9vf55"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.899044 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-db-sync-config-data\") pod \"barbican-db-sync-9vf55\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.899120 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrr57\" (UniqueName: \"kubernetes.io/projected/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-kube-api-access-xrr57\") pod \"barbican-db-sync-9vf55\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.899195 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-combined-ca-bundle\") pod \"barbican-db-sync-9vf55\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.911934 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-combined-ca-bundle\") pod \"barbican-db-sync-9vf55\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.915055 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-db-sync-config-data\") pod \"barbican-db-sync-9vf55\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.917965 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrr57\" (UniqueName: \"kubernetes.io/projected/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-kube-api-access-xrr57\") pod \"barbican-db-sync-9vf55\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.938041 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67cb7cd8c7-7q85d"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.942180 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.962920 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67cb7cd8c7-7q85d"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.976925 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.978363 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.985067 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6jj7z" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.985187 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.989194 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568fb5df6f-trz6g"] Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.999478 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 17 08:38:57 crc kubenswrapper[4966]: I1217 08:38:57.999685 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8vp7n" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.004977 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.007525 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.007991 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.018546 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568fb5df6f-trz6g"] Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.055064 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9vf55" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.105855 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.105951 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/03c06b3f-4fda-455c-9101-ee90d055ed88-horizon-secret-key\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.105982 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmm46\" (UniqueName: \"kubernetes.io/projected/b90ed9bf-cfa9-4410-9783-7ef1c162c876-kube-api-access-fmm46\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106012 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106039 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwbwh\" (UniqueName: \"kubernetes.io/projected/199ca5ed-4cae-4a6a-86c3-cf940a841b88-kube-api-access-gwbwh\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106070 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-config-data\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106098 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-svc\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106127 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-swift-storage-0\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106160 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-config\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106184 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-logs\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106204 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-nb\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106228 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-sb\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106259 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106295 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-scripts\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106331 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-config-data\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106355 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vkd2\" (UniqueName: \"kubernetes.io/projected/03c06b3f-4fda-455c-9101-ee90d055ed88-kube-api-access-8vkd2\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106444 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106477 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-scripts\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.106497 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03c06b3f-4fda-455c-9101-ee90d055ed88-logs\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.146597 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.148318 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.152116 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.152258 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.160158 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210072 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-logs\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210117 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210144 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-scripts\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210179 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-config-data\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210210 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vkd2\" (UniqueName: \"kubernetes.io/projected/03c06b3f-4fda-455c-9101-ee90d055ed88-kube-api-access-8vkd2\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210243 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210265 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw4mx\" (UniqueName: \"kubernetes.io/projected/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-kube-api-access-hw4mx\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210296 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210324 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210345 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-scripts\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210368 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210390 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03c06b3f-4fda-455c-9101-ee90d055ed88-logs\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210432 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210465 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmm46\" (UniqueName: \"kubernetes.io/projected/b90ed9bf-cfa9-4410-9783-7ef1c162c876-kube-api-access-fmm46\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210482 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/03c06b3f-4fda-455c-9101-ee90d055ed88-horizon-secret-key\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210506 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210526 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwbwh\" (UniqueName: \"kubernetes.io/projected/199ca5ed-4cae-4a6a-86c3-cf940a841b88-kube-api-access-gwbwh\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210546 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-config-data\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210568 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210586 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-svc\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210607 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-swift-storage-0\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210630 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-config\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210657 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-logs\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210687 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-nb\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210714 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-sb\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210744 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.210769 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.211564 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-scripts\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.211816 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03c06b3f-4fda-455c-9101-ee90d055ed88-logs\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.212559 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.212761 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-config-data\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.212771 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.215391 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-config\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.216066 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-svc\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.220268 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-swift-storage-0\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.224808 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-sb\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.225170 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-logs\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.225351 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-nb\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.226905 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/03c06b3f-4fda-455c-9101-ee90d055ed88-horizon-secret-key\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.239149 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.246094 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.250439 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vkd2\" (UniqueName: \"kubernetes.io/projected/03c06b3f-4fda-455c-9101-ee90d055ed88-kube-api-access-8vkd2\") pod \"horizon-67cb7cd8c7-7q85d\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.252507 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-scripts\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.255164 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-config-data\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.275758 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.285408 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwbwh\" (UniqueName: \"kubernetes.io/projected/199ca5ed-4cae-4a6a-86c3-cf940a841b88-kube-api-access-gwbwh\") pod \"dnsmasq-dns-568fb5df6f-trz6g\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.297170 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmm46\" (UniqueName: \"kubernetes.io/projected/b90ed9bf-cfa9-4410-9783-7ef1c162c876-kube-api-access-fmm46\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.312201 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.312280 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw4mx\" (UniqueName: \"kubernetes.io/projected/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-kube-api-access-hw4mx\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.312343 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.312366 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.312484 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.312586 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.312630 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-logs\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.312654 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.316807 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.325386 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.330187 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.338512 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.339328 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-logs\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.339798 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.359882 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.361587 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw4mx\" (UniqueName: \"kubernetes.io/projected/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-kube-api-access-hw4mx\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.383053 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.389427 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.402942 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-856cfb64ff-679wx"] Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.505463 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9bd96" event={"ID":"b7351028-bc40-4197-baa1-9fb0c525170d","Type":"ContainerStarted","Data":"be7c2be87eeece0352b016ee9dd70da9ddb5be531807821350ec778950dc323e"} Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.511493 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.517828 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-9bd96"] Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.521478 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-856cfb64ff-679wx" event={"ID":"e5a24615-9b89-4a84-b28b-cc5cf0d81682","Type":"ContainerStarted","Data":"9d6cc470f99ae56d7f89c7a1fe7f311f0d2d77c5f9bbea10e5fe1844b76914f3"} Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.567056 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.579978 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-69m6t"] Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.619305 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.642278 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rqvfb"] Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.817566 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fhj4g"] Dec 17 08:38:58 crc kubenswrapper[4966]: I1217 08:38:58.905133 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68d6578c4c-8tfk5"] Dec 17 08:38:58 crc kubenswrapper[4966]: W1217 08:38:58.930052 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb52f1e67_1d11_4eb6_bf67_ca0f3a5c1c62.slice/crio-1111b8b8fa3ac1c897a162a87a9db763626605c9e9ed11fd8deeb9c18c1a456a WatchSource:0}: Error finding container 1111b8b8fa3ac1c897a162a87a9db763626605c9e9ed11fd8deeb9c18c1a456a: Status 404 returned error can't find the container with id 1111b8b8fa3ac1c897a162a87a9db763626605c9e9ed11fd8deeb9c18c1a456a Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.266048 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67cb7cd8c7-7q85d"] Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.316397 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.373074 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6jj7z"] Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.400680 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9vf55"] Dec 17 08:38:59 crc kubenswrapper[4966]: W1217 08:38:59.447634 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeccb295b_14f1_43a9_98c8_673e2ce78add.slice/crio-9313cce76f78eee5bc25b7c1276d1086b88b7bd7c22f478cc5aecb8519f655ed WatchSource:0}: Error finding container 9313cce76f78eee5bc25b7c1276d1086b88b7bd7c22f478cc5aecb8519f655ed: Status 404 returned error can't find the container with id 9313cce76f78eee5bc25b7c1276d1086b88b7bd7c22f478cc5aecb8519f655ed Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.547585 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rqvfb" event={"ID":"3bca810d-b802-481a-8792-3be9b6a141bc","Type":"ContainerStarted","Data":"18376422af54c8345710b38bd7d0b1937e1efee1168738c20edf7fe0a44981f1"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.551095 4966 generic.go:334] "Generic (PLEG): container finished" podID="e5a24615-9b89-4a84-b28b-cc5cf0d81682" containerID="2aab64348abf97ee9ac5c0d5a0db58063b1c95b609f1d4eee37d3a366cf21b96" exitCode=0 Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.551134 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-856cfb64ff-679wx" event={"ID":"e5a24615-9b89-4a84-b28b-cc5cf0d81682","Type":"ContainerDied","Data":"2aab64348abf97ee9ac5c0d5a0db58063b1c95b609f1d4eee37d3a366cf21b96"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.556096 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6jj7z" event={"ID":"eccb295b-14f1-43a9-98c8-673e2ce78add","Type":"ContainerStarted","Data":"9313cce76f78eee5bc25b7c1276d1086b88b7bd7c22f478cc5aecb8519f655ed"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.596181 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68d6578c4c-8tfk5" event={"ID":"0c007a81-a0c1-43d8-8a77-62aebd2f188d","Type":"ContainerStarted","Data":"a3ab532487766ce991d3f4db25443c6704a80711ba3405f8ff813d3a1802e1e0"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.599710 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9vf55" event={"ID":"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a","Type":"ContainerStarted","Data":"69dacb6e019333d5f91c3ee578461bf033bacf167dbad0cbb83e6a155135d94b"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.617393 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-69m6t" event={"ID":"225a53d8-4292-436b-bb51-6527d51180c1","Type":"ContainerStarted","Data":"d235f2e117e97cc8381b5fedfe549c5806206e82a1a991bb6b0239c9826c9d51"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.617434 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-69m6t" event={"ID":"225a53d8-4292-436b-bb51-6527d51180c1","Type":"ContainerStarted","Data":"ddf6acb0727cc11232824aa279f7440a1f569566e1976aae4628684f83b4f566"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.630164 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fhj4g" event={"ID":"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62","Type":"ContainerStarted","Data":"9e2a12bca64d76f2997a3d267843d0bbdfedf734b641afb2cc96aaf0816e82a7"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.630220 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fhj4g" event={"ID":"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62","Type":"ContainerStarted","Data":"1111b8b8fa3ac1c897a162a87a9db763626605c9e9ed11fd8deeb9c18c1a456a"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.642717 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-69m6t" podStartSLOduration=3.642698354 podStartE2EDuration="3.642698354s" podCreationTimestamp="2025-12-17 08:38:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:59.641024248 +0000 UTC m=+1075.186094190" watchObservedRunningTime="2025-12-17 08:38:59.642698354 +0000 UTC m=+1075.187768296" Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.652148 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb7cd8c7-7q85d" event={"ID":"03c06b3f-4fda-455c-9101-ee90d055ed88","Type":"ContainerStarted","Data":"68776d2dd2a1e6367cded822199becccc9a3601e85a8306064ab0456ec7f7be1"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.656534 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerStarted","Data":"92447e11a1bfc272ca2bb89e988e45dc5b645f685826b7f261b07e71b5b73414"} Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.678084 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-fhj4g" podStartSLOduration=2.678065424 podStartE2EDuration="2.678065424s" podCreationTimestamp="2025-12-17 08:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:38:59.663411081 +0000 UTC m=+1075.208481023" watchObservedRunningTime="2025-12-17 08:38:59.678065424 +0000 UTC m=+1075.223135366" Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.736781 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:38:59 crc kubenswrapper[4966]: I1217 08:38:59.760555 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568fb5df6f-trz6g"] Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.070651 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.257944 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.348955 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67cb7cd8c7-7q85d"] Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.369643 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f89dfb859-v8j22"] Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.371027 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.506336 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f89dfb859-v8j22"] Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.507653 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-scripts\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.507721 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129625e4-4eef-46f8-861b-831451355683-logs\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.507780 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-config-data\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.507831 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/129625e4-4eef-46f8-861b-831451355683-horizon-secret-key\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.530587 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqxv4\" (UniqueName: \"kubernetes.io/projected/129625e4-4eef-46f8-861b-831451355683-kube-api-access-bqxv4\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.610105 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.633934 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqxv4\" (UniqueName: \"kubernetes.io/projected/129625e4-4eef-46f8-861b-831451355683-kube-api-access-bqxv4\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.634042 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-scripts\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.634180 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129625e4-4eef-46f8-861b-831451355683-logs\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.634359 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-config-data\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.634465 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/129625e4-4eef-46f8-861b-831451355683-horizon-secret-key\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.634793 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129625e4-4eef-46f8-861b-831451355683-logs\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.635087 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-scripts\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.636060 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-config-data\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.648475 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/129625e4-4eef-46f8-861b-831451355683-horizon-secret-key\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.667398 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.668738 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.700648 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqxv4\" (UniqueName: \"kubernetes.io/projected/129625e4-4eef-46f8-861b-831451355683-kube-api-access-bqxv4\") pod \"horizon-5f89dfb859-v8j22\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.707977 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.730122 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b90ed9bf-cfa9-4410-9783-7ef1c162c876","Type":"ContainerStarted","Data":"47e061b3288585a8fd8e7459cef4505967a744b973c5f821050c2d96d36b61d9"} Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.732641 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" event={"ID":"199ca5ed-4cae-4a6a-86c3-cf940a841b88","Type":"ContainerStarted","Data":"0ea705fe3f905a7f5c9c2ab549625001c1d12d75b5c98400ad978becf9d293cd"} Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.735427 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-sb\") pod \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.735477 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ls49\" (UniqueName: \"kubernetes.io/projected/e5a24615-9b89-4a84-b28b-cc5cf0d81682-kube-api-access-2ls49\") pod \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.735516 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-swift-storage-0\") pod \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.735596 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-svc\") pod \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.735616 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-config\") pod \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.735663 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-nb\") pod \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\" (UID: \"e5a24615-9b89-4a84-b28b-cc5cf0d81682\") " Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.763802 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5a24615-9b89-4a84-b28b-cc5cf0d81682-kube-api-access-2ls49" (OuterVolumeSpecName: "kube-api-access-2ls49") pod "e5a24615-9b89-4a84-b28b-cc5cf0d81682" (UID: "e5a24615-9b89-4a84-b28b-cc5cf0d81682"). InnerVolumeSpecName "kube-api-access-2ls49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.784219 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33","Type":"ContainerStarted","Data":"e50eb481c150acd0e2a9f00fbba49b6043758ae10d98b3e37d9600fb43a01bb3"} Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.785429 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e5a24615-9b89-4a84-b28b-cc5cf0d81682" (UID: "e5a24615-9b89-4a84-b28b-cc5cf0d81682"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.793921 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-856cfb64ff-679wx" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.794568 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-856cfb64ff-679wx" event={"ID":"e5a24615-9b89-4a84-b28b-cc5cf0d81682","Type":"ContainerDied","Data":"9d6cc470f99ae56d7f89c7a1fe7f311f0d2d77c5f9bbea10e5fe1844b76914f3"} Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.794608 4966 scope.go:117] "RemoveContainer" containerID="2aab64348abf97ee9ac5c0d5a0db58063b1c95b609f1d4eee37d3a366cf21b96" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.839173 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ls49\" (UniqueName: \"kubernetes.io/projected/e5a24615-9b89-4a84-b28b-cc5cf0d81682-kube-api-access-2ls49\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.839220 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.901153 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e5a24615-9b89-4a84-b28b-cc5cf0d81682" (UID: "e5a24615-9b89-4a84-b28b-cc5cf0d81682"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.910047 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e5a24615-9b89-4a84-b28b-cc5cf0d81682" (UID: "e5a24615-9b89-4a84-b28b-cc5cf0d81682"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.912398 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-config" (OuterVolumeSpecName: "config") pod "e5a24615-9b89-4a84-b28b-cc5cf0d81682" (UID: "e5a24615-9b89-4a84-b28b-cc5cf0d81682"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.914175 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e5a24615-9b89-4a84-b28b-cc5cf0d81682" (UID: "e5a24615-9b89-4a84-b28b-cc5cf0d81682"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.942338 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.942415 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.942429 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:00 crc kubenswrapper[4966]: I1217 08:39:00.942476 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5a24615-9b89-4a84-b28b-cc5cf0d81682-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:01 crc kubenswrapper[4966]: I1217 08:39:01.169094 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-856cfb64ff-679wx"] Dec 17 08:39:01 crc kubenswrapper[4966]: I1217 08:39:01.189825 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-856cfb64ff-679wx"] Dec 17 08:39:01 crc kubenswrapper[4966]: I1217 08:39:01.453048 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f89dfb859-v8j22"] Dec 17 08:39:01 crc kubenswrapper[4966]: I1217 08:39:01.879409 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f89dfb859-v8j22" event={"ID":"129625e4-4eef-46f8-861b-831451355683","Type":"ContainerStarted","Data":"2dd488b6a6dbfd7a5cb0294e6c995b6581844cbd8d4422168de836159d87af04"} Dec 17 08:39:01 crc kubenswrapper[4966]: I1217 08:39:01.902552 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b90ed9bf-cfa9-4410-9783-7ef1c162c876","Type":"ContainerStarted","Data":"9c14ed179cf5d0e21b47874bd77fdb37f9696a22277c7e74dc7b659e81a11798"} Dec 17 08:39:01 crc kubenswrapper[4966]: I1217 08:39:01.913159 4966 generic.go:334] "Generic (PLEG): container finished" podID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" containerID="6f0cdb24b2944687dac029575facceeb016def100367fbd904fe77f3c14e6ec1" exitCode=0 Dec 17 08:39:01 crc kubenswrapper[4966]: I1217 08:39:01.913192 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" event={"ID":"199ca5ed-4cae-4a6a-86c3-cf940a841b88","Type":"ContainerDied","Data":"6f0cdb24b2944687dac029575facceeb016def100367fbd904fe77f3c14e6ec1"} Dec 17 08:39:02 crc kubenswrapper[4966]: I1217 08:39:02.859939 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5a24615-9b89-4a84-b28b-cc5cf0d81682" path="/var/lib/kubelet/pods/e5a24615-9b89-4a84-b28b-cc5cf0d81682/volumes" Dec 17 08:39:02 crc kubenswrapper[4966]: I1217 08:39:02.950330 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" event={"ID":"199ca5ed-4cae-4a6a-86c3-cf940a841b88","Type":"ContainerStarted","Data":"6c8727b4b2f160093cb7b87a95b427dd52b19680f6d88b3dae14132e8c5bd476"} Dec 17 08:39:02 crc kubenswrapper[4966]: I1217 08:39:02.951398 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:39:02 crc kubenswrapper[4966]: I1217 08:39:02.983553 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" podStartSLOduration=5.983533945 podStartE2EDuration="5.983533945s" podCreationTimestamp="2025-12-17 08:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:02.974478307 +0000 UTC m=+1078.519548249" watchObservedRunningTime="2025-12-17 08:39:02.983533945 +0000 UTC m=+1078.528603887" Dec 17 08:39:03 crc kubenswrapper[4966]: I1217 08:39:03.073675 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33","Type":"ContainerStarted","Data":"3d75e4eaace730cdfca14100800b1f7cd7e07756b219c4f27803144c093bc3e1"} Dec 17 08:39:04 crc kubenswrapper[4966]: I1217 08:39:04.083858 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33","Type":"ContainerStarted","Data":"dcc0ca43aaebd497499db3c8dcaaf4d3e08926ba98c3fe2efc92140eefde3ae2"} Dec 17 08:39:04 crc kubenswrapper[4966]: I1217 08:39:04.084135 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerName="glance-log" containerID="cri-o://3d75e4eaace730cdfca14100800b1f7cd7e07756b219c4f27803144c093bc3e1" gracePeriod=30 Dec 17 08:39:04 crc kubenswrapper[4966]: I1217 08:39:04.084214 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerName="glance-httpd" containerID="cri-o://dcc0ca43aaebd497499db3c8dcaaf4d3e08926ba98c3fe2efc92140eefde3ae2" gracePeriod=30 Dec 17 08:39:04 crc kubenswrapper[4966]: I1217 08:39:04.090808 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerName="glance-log" containerID="cri-o://9c14ed179cf5d0e21b47874bd77fdb37f9696a22277c7e74dc7b659e81a11798" gracePeriod=30 Dec 17 08:39:04 crc kubenswrapper[4966]: I1217 08:39:04.090999 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b90ed9bf-cfa9-4410-9783-7ef1c162c876","Type":"ContainerStarted","Data":"72a5b43ec044c150683f6958968d9ae342846a65d4b7cdbbdc709d4806066704"} Dec 17 08:39:04 crc kubenswrapper[4966]: I1217 08:39:04.091061 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerName="glance-httpd" containerID="cri-o://72a5b43ec044c150683f6958968d9ae342846a65d4b7cdbbdc709d4806066704" gracePeriod=30 Dec 17 08:39:04 crc kubenswrapper[4966]: I1217 08:39:04.116626 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.116608483 podStartE2EDuration="7.116608483s" podCreationTimestamp="2025-12-17 08:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:04.103720619 +0000 UTC m=+1079.648790561" watchObservedRunningTime="2025-12-17 08:39:04.116608483 +0000 UTC m=+1079.661678415" Dec 17 08:39:04 crc kubenswrapper[4966]: I1217 08:39:04.867625 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.867608657 podStartE2EDuration="7.867608657s" podCreationTimestamp="2025-12-17 08:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:04.133020842 +0000 UTC m=+1079.678090804" watchObservedRunningTime="2025-12-17 08:39:04.867608657 +0000 UTC m=+1080.412678599" Dec 17 08:39:05 crc kubenswrapper[4966]: I1217 08:39:05.135992 4966 generic.go:334] "Generic (PLEG): container finished" podID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerID="dcc0ca43aaebd497499db3c8dcaaf4d3e08926ba98c3fe2efc92140eefde3ae2" exitCode=0 Dec 17 08:39:05 crc kubenswrapper[4966]: I1217 08:39:05.136024 4966 generic.go:334] "Generic (PLEG): container finished" podID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerID="3d75e4eaace730cdfca14100800b1f7cd7e07756b219c4f27803144c093bc3e1" exitCode=143 Dec 17 08:39:05 crc kubenswrapper[4966]: I1217 08:39:05.136072 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33","Type":"ContainerDied","Data":"dcc0ca43aaebd497499db3c8dcaaf4d3e08926ba98c3fe2efc92140eefde3ae2"} Dec 17 08:39:05 crc kubenswrapper[4966]: I1217 08:39:05.136104 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33","Type":"ContainerDied","Data":"3d75e4eaace730cdfca14100800b1f7cd7e07756b219c4f27803144c093bc3e1"} Dec 17 08:39:05 crc kubenswrapper[4966]: I1217 08:39:05.147804 4966 generic.go:334] "Generic (PLEG): container finished" podID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerID="72a5b43ec044c150683f6958968d9ae342846a65d4b7cdbbdc709d4806066704" exitCode=0 Dec 17 08:39:05 crc kubenswrapper[4966]: I1217 08:39:05.147836 4966 generic.go:334] "Generic (PLEG): container finished" podID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerID="9c14ed179cf5d0e21b47874bd77fdb37f9696a22277c7e74dc7b659e81a11798" exitCode=143 Dec 17 08:39:05 crc kubenswrapper[4966]: I1217 08:39:05.148190 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b90ed9bf-cfa9-4410-9783-7ef1c162c876","Type":"ContainerDied","Data":"72a5b43ec044c150683f6958968d9ae342846a65d4b7cdbbdc709d4806066704"} Dec 17 08:39:05 crc kubenswrapper[4966]: I1217 08:39:05.148299 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b90ed9bf-cfa9-4410-9783-7ef1c162c876","Type":"ContainerDied","Data":"9c14ed179cf5d0e21b47874bd77fdb37f9696a22277c7e74dc7b659e81a11798"} Dec 17 08:39:06 crc kubenswrapper[4966]: I1217 08:39:06.954063 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68d6578c4c-8tfk5"] Dec 17 08:39:06 crc kubenswrapper[4966]: I1217 08:39:06.988930 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-fbfdd886-69frk"] Dec 17 08:39:06 crc kubenswrapper[4966]: E1217 08:39:06.989367 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5a24615-9b89-4a84-b28b-cc5cf0d81682" containerName="init" Dec 17 08:39:06 crc kubenswrapper[4966]: I1217 08:39:06.989383 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5a24615-9b89-4a84-b28b-cc5cf0d81682" containerName="init" Dec 17 08:39:06 crc kubenswrapper[4966]: I1217 08:39:06.989622 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5a24615-9b89-4a84-b28b-cc5cf0d81682" containerName="init" Dec 17 08:39:06 crc kubenswrapper[4966]: I1217 08:39:06.993622 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.005495 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.006932 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-fbfdd886-69frk"] Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.081907 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f89dfb859-v8j22"] Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.124862 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-secret-key\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.124919 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xvb9\" (UniqueName: \"kubernetes.io/projected/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-kube-api-access-4xvb9\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.124942 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-config-data\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.125017 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-combined-ca-bundle\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.125035 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-logs\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.125062 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-scripts\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.125115 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-tls-certs\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.139509 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-55d57cbbcb-2tmk5"] Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.148884 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.167903 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55d57cbbcb-2tmk5"] Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.228277 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-tls-certs\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.228396 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-secret-key\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.228418 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xvb9\" (UniqueName: \"kubernetes.io/projected/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-kube-api-access-4xvb9\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.228443 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-config-data\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.228498 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-combined-ca-bundle\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.228523 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-logs\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.228560 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-scripts\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.230082 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-scripts\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.232476 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-logs\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.233934 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-config-data\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.240718 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-secret-key\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.244726 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-combined-ca-bundle\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.247484 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-tls-certs\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.250732 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xvb9\" (UniqueName: \"kubernetes.io/projected/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-kube-api-access-4xvb9\") pod \"horizon-fbfdd886-69frk\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.329698 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3ade9e0-df6d-4b67-a0a3-727dc696a358-config-data\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.330096 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3ade9e0-df6d-4b67-a0a3-727dc696a358-scripts\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.330139 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f3ade9e0-df6d-4b67-a0a3-727dc696a358-horizon-secret-key\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.330164 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3ade9e0-df6d-4b67-a0a3-727dc696a358-logs\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.330178 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ade9e0-df6d-4b67-a0a3-727dc696a358-combined-ca-bundle\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.330223 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mx89\" (UniqueName: \"kubernetes.io/projected/f3ade9e0-df6d-4b67-a0a3-727dc696a358-kube-api-access-7mx89\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.330293 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ade9e0-df6d-4b67-a0a3-727dc696a358-horizon-tls-certs\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.345471 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.432771 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mx89\" (UniqueName: \"kubernetes.io/projected/f3ade9e0-df6d-4b67-a0a3-727dc696a358-kube-api-access-7mx89\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.432904 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ade9e0-df6d-4b67-a0a3-727dc696a358-horizon-tls-certs\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.432931 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3ade9e0-df6d-4b67-a0a3-727dc696a358-config-data\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.432991 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3ade9e0-df6d-4b67-a0a3-727dc696a358-scripts\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.433021 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f3ade9e0-df6d-4b67-a0a3-727dc696a358-horizon-secret-key\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.433038 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3ade9e0-df6d-4b67-a0a3-727dc696a358-logs\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.433054 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ade9e0-df6d-4b67-a0a3-727dc696a358-combined-ca-bundle\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.434154 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f3ade9e0-df6d-4b67-a0a3-727dc696a358-config-data\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.434694 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3ade9e0-df6d-4b67-a0a3-727dc696a358-scripts\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.434919 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3ade9e0-df6d-4b67-a0a3-727dc696a358-logs\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.439681 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ade9e0-df6d-4b67-a0a3-727dc696a358-combined-ca-bundle\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.440071 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f3ade9e0-df6d-4b67-a0a3-727dc696a358-horizon-secret-key\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.441560 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ade9e0-df6d-4b67-a0a3-727dc696a358-horizon-tls-certs\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.476479 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mx89\" (UniqueName: \"kubernetes.io/projected/f3ade9e0-df6d-4b67-a0a3-727dc696a358-kube-api-access-7mx89\") pod \"horizon-55d57cbbcb-2tmk5\" (UID: \"f3ade9e0-df6d-4b67-a0a3-727dc696a358\") " pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:07 crc kubenswrapper[4966]: I1217 08:39:07.767851 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:08 crc kubenswrapper[4966]: I1217 08:39:08.187985 4966 generic.go:334] "Generic (PLEG): container finished" podID="225a53d8-4292-436b-bb51-6527d51180c1" containerID="d235f2e117e97cc8381b5fedfe549c5806206e82a1a991bb6b0239c9826c9d51" exitCode=0 Dec 17 08:39:08 crc kubenswrapper[4966]: I1217 08:39:08.188037 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-69m6t" event={"ID":"225a53d8-4292-436b-bb51-6527d51180c1","Type":"ContainerDied","Data":"d235f2e117e97cc8381b5fedfe549c5806206e82a1a991bb6b0239c9826c9d51"} Dec 17 08:39:08 crc kubenswrapper[4966]: I1217 08:39:08.570530 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:39:08 crc kubenswrapper[4966]: I1217 08:39:08.624804 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b75d776ff-f25pq"] Dec 17 08:39:08 crc kubenswrapper[4966]: I1217 08:39:08.625087 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" containerID="cri-o://f54a4f58f7bb67f9d87e7addb7489a5b71d139a3f9c89fc2c158fe0bbd974742" gracePeriod=10 Dec 17 08:39:09 crc kubenswrapper[4966]: I1217 08:39:09.215032 4966 generic.go:334] "Generic (PLEG): container finished" podID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerID="f54a4f58f7bb67f9d87e7addb7489a5b71d139a3f9c89fc2c158fe0bbd974742" exitCode=0 Dec 17 08:39:09 crc kubenswrapper[4966]: I1217 08:39:09.215194 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" event={"ID":"28af8586-3f33-42a1-813b-a1382cde5ec9","Type":"ContainerDied","Data":"f54a4f58f7bb67f9d87e7addb7489a5b71d139a3f9c89fc2c158fe0bbd974742"} Dec 17 08:39:09 crc kubenswrapper[4966]: I1217 08:39:09.410574 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: connect: connection refused" Dec 17 08:39:14 crc kubenswrapper[4966]: I1217 08:39:14.410742 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: connect: connection refused" Dec 17 08:39:17 crc kubenswrapper[4966]: E1217 08:39:17.083195 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-placement-api:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:17 crc kubenswrapper[4966]: E1217 08:39:17.083512 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-placement-api:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:17 crc kubenswrapper[4966]: E1217 08:39:17.083646 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-placement-api:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lp27h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-6jj7z_openstack(eccb295b-14f1-43a9-98c8-673e2ce78add): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:39:17 crc kubenswrapper[4966]: E1217 08:39:17.084828 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-6jj7z" podUID="eccb295b-14f1-43a9-98c8-673e2ce78add" Dec 17 08:39:17 crc kubenswrapper[4966]: E1217 08:39:17.292127 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.156:5001/podified-antelope-centos9/openstack-placement-api:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/placement-db-sync-6jj7z" podUID="eccb295b-14f1-43a9-98c8-673e2ce78add" Dec 17 08:39:19 crc kubenswrapper[4966]: I1217 08:39:19.411444 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: connect: connection refused" Dec 17 08:39:19 crc kubenswrapper[4966]: I1217 08:39:19.411948 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:39:23 crc kubenswrapper[4966]: E1217 08:39:23.561811 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-horizon:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:23 crc kubenswrapper[4966]: E1217 08:39:23.562977 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-horizon:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:23 crc kubenswrapper[4966]: E1217 08:39:23.563166 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-horizon:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc8h56h554h569h8ch8ch556h5b7h586h55dh5bch5d4h5h5dch654h6hf4hc7h5f9h77h58fh78h5bhc8hddh64h568h5d4h5dfh5f8h57fh576q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vqrgz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-68d6578c4c-8tfk5_openstack(0c007a81-a0c1-43d8-8a77-62aebd2f188d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:39:23 crc kubenswrapper[4966]: E1217 08:39:23.566242 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.156:5001/podified-antelope-centos9/openstack-horizon:c3923531bcda0b0811b2d5053f189beb\\\"\"]" pod="openstack/horizon-68d6578c4c-8tfk5" podUID="0c007a81-a0c1-43d8-8a77-62aebd2f188d" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.655011 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.685978 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.690087 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.859555 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-public-tls-certs\") pod \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.859636 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-combined-ca-bundle\") pod \"225a53d8-4292-436b-bb51-6527d51180c1\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.859656 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-scripts\") pod \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.860881 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-logs" (OuterVolumeSpecName: "logs") pod "b90ed9bf-cfa9-4410-9783-7ef1c162c876" (UID: "b90ed9bf-cfa9-4410-9783-7ef1c162c876"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.860919 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-logs\") pod \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.860944 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-scripts\") pod \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.861011 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-combined-ca-bundle\") pod \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.861027 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.861046 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-httpd-run\") pod \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.861090 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-combined-ca-bundle\") pod \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.861213 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmm46\" (UniqueName: \"kubernetes.io/projected/b90ed9bf-cfa9-4410-9783-7ef1c162c876-kube-api-access-fmm46\") pod \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.861248 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-scripts\") pod \"225a53d8-4292-436b-bb51-6527d51180c1\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.861264 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-credential-keys\") pod \"225a53d8-4292-436b-bb51-6527d51180c1\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.861512 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b90ed9bf-cfa9-4410-9783-7ef1c162c876" (UID: "b90ed9bf-cfa9-4410-9783-7ef1c162c876"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.861986 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-httpd-run\") pod \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.862095 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29mzh\" (UniqueName: \"kubernetes.io/projected/225a53d8-4292-436b-bb51-6527d51180c1-kube-api-access-29mzh\") pod \"225a53d8-4292-436b-bb51-6527d51180c1\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.862117 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-config-data\") pod \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\" (UID: \"b90ed9bf-cfa9-4410-9783-7ef1c162c876\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.862163 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.862189 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-config-data\") pod \"225a53d8-4292-436b-bb51-6527d51180c1\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.862255 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-internal-tls-certs\") pod \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.862294 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-fernet-keys\") pod \"225a53d8-4292-436b-bb51-6527d51180c1\" (UID: \"225a53d8-4292-436b-bb51-6527d51180c1\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.862337 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-config-data\") pod \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.862364 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw4mx\" (UniqueName: \"kubernetes.io/projected/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-kube-api-access-hw4mx\") pod \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.862906 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-logs\") pod \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\" (UID: \"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33\") " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.863188 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" (UID: "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.863593 4966 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.863607 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.863614 4966 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b90ed9bf-cfa9-4410-9783-7ef1c162c876-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.865381 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-scripts" (OuterVolumeSpecName: "scripts") pod "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" (UID: "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.866673 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "b90ed9bf-cfa9-4410-9783-7ef1c162c876" (UID: "b90ed9bf-cfa9-4410-9783-7ef1c162c876"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.866953 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-logs" (OuterVolumeSpecName: "logs") pod "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" (UID: "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.867151 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-scripts" (OuterVolumeSpecName: "scripts") pod "b90ed9bf-cfa9-4410-9783-7ef1c162c876" (UID: "b90ed9bf-cfa9-4410-9783-7ef1c162c876"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.867769 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "225a53d8-4292-436b-bb51-6527d51180c1" (UID: "225a53d8-4292-436b-bb51-6527d51180c1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.870790 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b90ed9bf-cfa9-4410-9783-7ef1c162c876-kube-api-access-fmm46" (OuterVolumeSpecName: "kube-api-access-fmm46") pod "b90ed9bf-cfa9-4410-9783-7ef1c162c876" (UID: "b90ed9bf-cfa9-4410-9783-7ef1c162c876"). InnerVolumeSpecName "kube-api-access-fmm46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.873682 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-scripts" (OuterVolumeSpecName: "scripts") pod "225a53d8-4292-436b-bb51-6527d51180c1" (UID: "225a53d8-4292-436b-bb51-6527d51180c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.873790 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/225a53d8-4292-436b-bb51-6527d51180c1-kube-api-access-29mzh" (OuterVolumeSpecName: "kube-api-access-29mzh") pod "225a53d8-4292-436b-bb51-6527d51180c1" (UID: "225a53d8-4292-436b-bb51-6527d51180c1"). InnerVolumeSpecName "kube-api-access-29mzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.887289 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-kube-api-access-hw4mx" (OuterVolumeSpecName: "kube-api-access-hw4mx") pod "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" (UID: "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33"). InnerVolumeSpecName "kube-api-access-hw4mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.890480 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" (UID: "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.890675 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "225a53d8-4292-436b-bb51-6527d51180c1" (UID: "225a53d8-4292-436b-bb51-6527d51180c1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.896757 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "225a53d8-4292-436b-bb51-6527d51180c1" (UID: "225a53d8-4292-436b-bb51-6527d51180c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.898514 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" (UID: "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.923660 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-config-data" (OuterVolumeSpecName: "config-data") pod "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" (UID: "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.926158 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" (UID: "0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.932854 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b90ed9bf-cfa9-4410-9783-7ef1c162c876" (UID: "b90ed9bf-cfa9-4410-9783-7ef1c162c876"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.938862 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-config-data" (OuterVolumeSpecName: "config-data") pod "225a53d8-4292-436b-bb51-6527d51180c1" (UID: "225a53d8-4292-436b-bb51-6527d51180c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.953242 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b90ed9bf-cfa9-4410-9783-7ef1c162c876" (UID: "b90ed9bf-cfa9-4410-9783-7ef1c162c876"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964139 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29mzh\" (UniqueName: \"kubernetes.io/projected/225a53d8-4292-436b-bb51-6527d51180c1-kube-api-access-29mzh\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964183 4966 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964194 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964206 4966 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964215 4966 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964222 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964231 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw4mx\" (UniqueName: \"kubernetes.io/projected/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-kube-api-access-hw4mx\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964240 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964248 4966 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964256 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964264 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964272 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964280 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964298 4966 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964309 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964317 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmm46\" (UniqueName: \"kubernetes.io/projected/b90ed9bf-cfa9-4410-9783-7ef1c162c876-kube-api-access-fmm46\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964325 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.964332 4966 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/225a53d8-4292-436b-bb51-6527d51180c1-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.970249 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-config-data" (OuterVolumeSpecName: "config-data") pod "b90ed9bf-cfa9-4410-9783-7ef1c162c876" (UID: "b90ed9bf-cfa9-4410-9783-7ef1c162c876"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.982169 4966 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 17 08:39:23 crc kubenswrapper[4966]: I1217 08:39:23.987482 4966 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.066241 4966 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.066275 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90ed9bf-cfa9-4410-9783-7ef1c162c876-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.066285 4966 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.357811 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b90ed9bf-cfa9-4410-9783-7ef1c162c876","Type":"ContainerDied","Data":"47e061b3288585a8fd8e7459cef4505967a744b973c5f821050c2d96d36b61d9"} Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.358181 4966 scope.go:117] "RemoveContainer" containerID="72a5b43ec044c150683f6958968d9ae342846a65d4b7cdbbdc709d4806066704" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.357891 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.360664 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-69m6t" event={"ID":"225a53d8-4292-436b-bb51-6527d51180c1","Type":"ContainerDied","Data":"ddf6acb0727cc11232824aa279f7440a1f569566e1976aae4628684f83b4f566"} Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.360711 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-69m6t" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.360721 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddf6acb0727cc11232824aa279f7440a1f569566e1976aae4628684f83b4f566" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.363771 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.374690 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33","Type":"ContainerDied","Data":"e50eb481c150acd0e2a9f00fbba49b6043758ae10d98b3e37d9600fb43a01bb3"} Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.452205 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.461186 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.474216 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.511020 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.522427 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:39:24 crc kubenswrapper[4966]: E1217 08:39:24.523044 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerName="glance-httpd" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523068 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerName="glance-httpd" Dec 17 08:39:24 crc kubenswrapper[4966]: E1217 08:39:24.523094 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerName="glance-log" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523104 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerName="glance-log" Dec 17 08:39:24 crc kubenswrapper[4966]: E1217 08:39:24.523118 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="225a53d8-4292-436b-bb51-6527d51180c1" containerName="keystone-bootstrap" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523127 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="225a53d8-4292-436b-bb51-6527d51180c1" containerName="keystone-bootstrap" Dec 17 08:39:24 crc kubenswrapper[4966]: E1217 08:39:24.523150 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerName="glance-httpd" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523158 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerName="glance-httpd" Dec 17 08:39:24 crc kubenswrapper[4966]: E1217 08:39:24.523176 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerName="glance-log" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523183 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerName="glance-log" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523408 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerName="glance-log" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523425 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerName="glance-httpd" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523439 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" containerName="glance-log" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523457 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" containerName="glance-httpd" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.523474 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="225a53d8-4292-436b-bb51-6527d51180c1" containerName="keystone-bootstrap" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.524790 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.527201 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.527234 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.527476 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8vp7n" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.527707 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.530734 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.541735 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.543534 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.546720 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.547239 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.553649 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.600790 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-config-data\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.601771 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.601862 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w427r\" (UniqueName: \"kubernetes.io/projected/d7a7b745-08ce-4be8-b81b-82f632be09a3-kube-api-access-w427r\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.602239 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.602327 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.602400 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.602486 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-scripts\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.602635 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-logs\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704181 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w427r\" (UniqueName: \"kubernetes.io/projected/d7a7b745-08ce-4be8-b81b-82f632be09a3-kube-api-access-w427r\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704234 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704261 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704288 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704314 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704357 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-scripts\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704389 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-scripts\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704406 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-config-data\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704422 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704460 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbhl4\" (UniqueName: \"kubernetes.io/projected/073e8454-1a1f-4a48-9516-070e0fddc046-kube-api-access-hbhl4\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704479 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704514 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-logs\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704538 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704554 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-logs\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704579 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-config-data\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.704596 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.705356 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.705789 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.705828 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-logs\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.794661 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w427r\" (UniqueName: \"kubernetes.io/projected/d7a7b745-08ce-4be8-b81b-82f632be09a3-kube-api-access-w427r\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.799413 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-config-data\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.800761 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.819643 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.846208 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.846261 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-logs\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.846324 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.846429 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-scripts\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.846450 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-config-data\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.846471 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.846513 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbhl4\" (UniqueName: \"kubernetes.io/projected/073e8454-1a1f-4a48-9516-070e0fddc046-kube-api-access-hbhl4\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.846545 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.864489 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-scripts\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.938496 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.949284 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:24 crc kubenswrapper[4966]: I1217 08:39:24.997136 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.006387 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-logs\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.009318 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-config-data\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.011469 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.023408 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " pod="openstack/glance-default-external-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.026029 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33" path="/var/lib/kubelet/pods/0aaf8d87-8cf8-4a82-9702-cc1dc5ff2f33/volumes" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.026658 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b90ed9bf-cfa9-4410-9783-7ef1c162c876" path="/var/lib/kubelet/pods/b90ed9bf-cfa9-4410-9783-7ef1c162c876/volumes" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.028394 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-scripts\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.029239 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbhl4\" (UniqueName: \"kubernetes.io/projected/073e8454-1a1f-4a48-9516-070e0fddc046-kube-api-access-hbhl4\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.063385 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.097963 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-gkjk7"] Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.099249 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.104398 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.104702 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.104817 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.104940 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-czvn7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.114587 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.148750 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gkjk7"] Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.160403 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.180083 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-69m6t"] Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.191537 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.203400 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-69m6t"] Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.249338 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-fernet-keys\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.249415 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-credential-keys\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.249459 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-scripts\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.249489 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-combined-ca-bundle\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.249586 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-config-data\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.249633 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb774\" (UniqueName: \"kubernetes.io/projected/95497f6e-9ff6-457b-a47a-3747cc4ddc69-kube-api-access-zb774\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.350969 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-fernet-keys\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.351056 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-credential-keys\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.351093 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-scripts\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.351133 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-combined-ca-bundle\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.351223 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-config-data\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.351259 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb774\" (UniqueName: \"kubernetes.io/projected/95497f6e-9ff6-457b-a47a-3747cc4ddc69-kube-api-access-zb774\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.354524 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-combined-ca-bundle\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.354856 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-scripts\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.355123 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-fernet-keys\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.355336 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-config-data\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.356652 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-credential-keys\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.369029 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb774\" (UniqueName: \"kubernetes.io/projected/95497f6e-9ff6-457b-a47a-3747cc4ddc69-kube-api-access-zb774\") pod \"keystone-bootstrap-gkjk7\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:25 crc kubenswrapper[4966]: I1217 08:39:25.466769 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:26 crc kubenswrapper[4966]: I1217 08:39:26.844151 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="225a53d8-4292-436b-bb51-6527d51180c1" path="/var/lib/kubelet/pods/225a53d8-4292-436b-bb51-6527d51180c1/volumes" Dec 17 08:39:29 crc kubenswrapper[4966]: I1217 08:39:29.411149 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Dec 17 08:39:30 crc kubenswrapper[4966]: I1217 08:39:30.416277 4966 generic.go:334] "Generic (PLEG): container finished" podID="b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62" containerID="9e2a12bca64d76f2997a3d267843d0bbdfedf734b641afb2cc96aaf0816e82a7" exitCode=0 Dec 17 08:39:30 crc kubenswrapper[4966]: I1217 08:39:30.416372 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fhj4g" event={"ID":"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62","Type":"ContainerDied","Data":"9e2a12bca64d76f2997a3d267843d0bbdfedf734b641afb2cc96aaf0816e82a7"} Dec 17 08:39:34 crc kubenswrapper[4966]: E1217 08:39:34.395157 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-barbican-api:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:34 crc kubenswrapper[4966]: E1217 08:39:34.395855 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-barbican-api:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:34 crc kubenswrapper[4966]: E1217 08:39:34.396104 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-barbican-api:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xrr57,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-9vf55_openstack(2ceb72da-ea40-4c24-8b26-94b10ba9cc6a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:39:34 crc kubenswrapper[4966]: E1217 08:39:34.397375 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-9vf55" podUID="2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" Dec 17 08:39:34 crc kubenswrapper[4966]: I1217 08:39:34.412144 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Dec 17 08:39:34 crc kubenswrapper[4966]: E1217 08:39:34.461831 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.156:5001/podified-antelope-centos9/openstack-barbican-api:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/barbican-db-sync-9vf55" podUID="2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" Dec 17 08:39:34 crc kubenswrapper[4966]: E1217 08:39:34.677317 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-heat-engine:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:34 crc kubenswrapper[4966]: E1217 08:39:34.677628 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-heat-engine:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:34 crc kubenswrapper[4966]: E1217 08:39:34.678037 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-heat-engine:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pbq6s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-9bd96_openstack(b7351028-bc40-4197-baa1-9fb0c525170d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:39:34 crc kubenswrapper[4966]: E1217 08:39:34.679289 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-9bd96" podUID="b7351028-bc40-4197-baa1-9fb0c525170d" Dec 17 08:39:34 crc kubenswrapper[4966]: I1217 08:39:34.791887 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:39:34 crc kubenswrapper[4966]: I1217 08:39:34.800312 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:39:34 crc kubenswrapper[4966]: I1217 08:39:34.808691 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.015517 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw457\" (UniqueName: \"kubernetes.io/projected/28af8586-3f33-42a1-813b-a1382cde5ec9-kube-api-access-kw457\") pod \"28af8586-3f33-42a1-813b-a1382cde5ec9\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.015898 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-config-data\") pod \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.015936 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-config\") pod \"28af8586-3f33-42a1-813b-a1382cde5ec9\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.016679 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-config-data" (OuterVolumeSpecName: "config-data") pod "0c007a81-a0c1-43d8-8a77-62aebd2f188d" (UID: "0c007a81-a0c1-43d8-8a77-62aebd2f188d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.016744 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-combined-ca-bundle\") pod \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.016784 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqrgz\" (UniqueName: \"kubernetes.io/projected/0c007a81-a0c1-43d8-8a77-62aebd2f188d-kube-api-access-vqrgz\") pod \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.016856 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-swift-storage-0\") pod \"28af8586-3f33-42a1-813b-a1382cde5ec9\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.016918 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c007a81-a0c1-43d8-8a77-62aebd2f188d-horizon-secret-key\") pod \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.016945 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlm7l\" (UniqueName: \"kubernetes.io/projected/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-kube-api-access-dlm7l\") pod \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.016959 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-nb\") pod \"28af8586-3f33-42a1-813b-a1382cde5ec9\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.017019 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-scripts\") pod \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.017050 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-svc\") pod \"28af8586-3f33-42a1-813b-a1382cde5ec9\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.017091 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c007a81-a0c1-43d8-8a77-62aebd2f188d-logs\") pod \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\" (UID: \"0c007a81-a0c1-43d8-8a77-62aebd2f188d\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.017115 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-sb\") pod \"28af8586-3f33-42a1-813b-a1382cde5ec9\" (UID: \"28af8586-3f33-42a1-813b-a1382cde5ec9\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.017158 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-config\") pod \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\" (UID: \"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62\") " Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.017650 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.018683 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-scripts" (OuterVolumeSpecName: "scripts") pod "0c007a81-a0c1-43d8-8a77-62aebd2f188d" (UID: "0c007a81-a0c1-43d8-8a77-62aebd2f188d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.019724 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28af8586-3f33-42a1-813b-a1382cde5ec9-kube-api-access-kw457" (OuterVolumeSpecName: "kube-api-access-kw457") pod "28af8586-3f33-42a1-813b-a1382cde5ec9" (UID: "28af8586-3f33-42a1-813b-a1382cde5ec9"). InnerVolumeSpecName "kube-api-access-kw457". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.035671 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c007a81-a0c1-43d8-8a77-62aebd2f188d-logs" (OuterVolumeSpecName: "logs") pod "0c007a81-a0c1-43d8-8a77-62aebd2f188d" (UID: "0c007a81-a0c1-43d8-8a77-62aebd2f188d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.039308 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c007a81-a0c1-43d8-8a77-62aebd2f188d-kube-api-access-vqrgz" (OuterVolumeSpecName: "kube-api-access-vqrgz") pod "0c007a81-a0c1-43d8-8a77-62aebd2f188d" (UID: "0c007a81-a0c1-43d8-8a77-62aebd2f188d"). InnerVolumeSpecName "kube-api-access-vqrgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.077658 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-kube-api-access-dlm7l" (OuterVolumeSpecName: "kube-api-access-dlm7l") pod "b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62" (UID: "b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62"). InnerVolumeSpecName "kube-api-access-dlm7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.087051 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c007a81-a0c1-43d8-8a77-62aebd2f188d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0c007a81-a0c1-43d8-8a77-62aebd2f188d" (UID: "0c007a81-a0c1-43d8-8a77-62aebd2f188d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.126333 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-fbfdd886-69frk"] Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.128885 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw457\" (UniqueName: \"kubernetes.io/projected/28af8586-3f33-42a1-813b-a1382cde5ec9-kube-api-access-kw457\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.128909 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqrgz\" (UniqueName: \"kubernetes.io/projected/0c007a81-a0c1-43d8-8a77-62aebd2f188d-kube-api-access-vqrgz\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.128919 4966 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0c007a81-a0c1-43d8-8a77-62aebd2f188d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.128939 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlm7l\" (UniqueName: \"kubernetes.io/projected/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-kube-api-access-dlm7l\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.128949 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0c007a81-a0c1-43d8-8a77-62aebd2f188d-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.128958 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c007a81-a0c1-43d8-8a77-62aebd2f188d-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.132177 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62" (UID: "b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.139011 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-config" (OuterVolumeSpecName: "config") pod "b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62" (UID: "b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.158538 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "28af8586-3f33-42a1-813b-a1382cde5ec9" (UID: "28af8586-3f33-42a1-813b-a1382cde5ec9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.165413 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-config" (OuterVolumeSpecName: "config") pod "28af8586-3f33-42a1-813b-a1382cde5ec9" (UID: "28af8586-3f33-42a1-813b-a1382cde5ec9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.169990 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "28af8586-3f33-42a1-813b-a1382cde5ec9" (UID: "28af8586-3f33-42a1-813b-a1382cde5ec9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.172976 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "28af8586-3f33-42a1-813b-a1382cde5ec9" (UID: "28af8586-3f33-42a1-813b-a1382cde5ec9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.173346 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "28af8586-3f33-42a1-813b-a1382cde5ec9" (UID: "28af8586-3f33-42a1-813b-a1382cde5ec9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.230937 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.230965 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.230976 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.230984 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.230993 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.231001 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.231009 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28af8586-3f33-42a1-813b-a1382cde5ec9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.470479 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" event={"ID":"28af8586-3f33-42a1-813b-a1382cde5ec9","Type":"ContainerDied","Data":"a01a4e8019ab88346ae2f90c36181219f46649c45a8b98cae5c1eadc7365bcec"} Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.470587 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.472300 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68d6578c4c-8tfk5" event={"ID":"0c007a81-a0c1-43d8-8a77-62aebd2f188d","Type":"ContainerDied","Data":"a3ab532487766ce991d3f4db25443c6704a80711ba3405f8ff813d3a1802e1e0"} Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.472341 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68d6578c4c-8tfk5" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.475862 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fhj4g" event={"ID":"b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62","Type":"ContainerDied","Data":"1111b8b8fa3ac1c897a162a87a9db763626605c9e9ed11fd8deeb9c18c1a456a"} Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.475939 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1111b8b8fa3ac1c897a162a87a9db763626605c9e9ed11fd8deeb9c18c1a456a" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.475953 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fhj4g" Dec 17 08:39:35 crc kubenswrapper[4966]: E1217 08:39:35.486197 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.156:5001/podified-antelope-centos9/openstack-heat-engine:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/heat-db-sync-9bd96" podUID="b7351028-bc40-4197-baa1-9fb0c525170d" Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.562407 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68d6578c4c-8tfk5"] Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.575121 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68d6578c4c-8tfk5"] Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.585799 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b75d776ff-f25pq"] Dec 17 08:39:35 crc kubenswrapper[4966]: I1217 08:39:35.594181 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b75d776ff-f25pq"] Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.209842 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f4484465c-9cs9d"] Dec 17 08:39:36 crc kubenswrapper[4966]: E1217 08:39:36.210433 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.210444 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" Dec 17 08:39:36 crc kubenswrapper[4966]: E1217 08:39:36.210461 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="init" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.210467 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="init" Dec 17 08:39:36 crc kubenswrapper[4966]: E1217 08:39:36.210481 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62" containerName="neutron-db-sync" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.210488 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62" containerName="neutron-db-sync" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.210667 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.210683 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62" containerName="neutron-db-sync" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.211650 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.248690 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f4484465c-9cs9d"] Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.361003 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-sb\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.361085 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-664fk\" (UniqueName: \"kubernetes.io/projected/cff2f826-ba4a-4fe1-9406-bfdb021aea19-kube-api-access-664fk\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.361146 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-config\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.361182 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-swift-storage-0\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.361203 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-nb\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.361276 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-svc\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.379916 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c574bf78b-4thfs"] Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.383735 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.392179 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.392682 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.392796 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.392913 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-bx4bj" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.400075 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c574bf78b-4thfs"] Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.462922 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-svc\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.463010 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-sb\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.463050 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-664fk\" (UniqueName: \"kubernetes.io/projected/cff2f826-ba4a-4fe1-9406-bfdb021aea19-kube-api-access-664fk\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.463088 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-config\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.463114 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-swift-storage-0\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.463135 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-nb\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.463949 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-nb\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.464448 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-svc\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.464932 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-sb\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.465672 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-config\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.466171 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-swift-storage-0\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.481279 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-664fk\" (UniqueName: \"kubernetes.io/projected/cff2f826-ba4a-4fe1-9406-bfdb021aea19-kube-api-access-664fk\") pod \"dnsmasq-dns-5f4484465c-9cs9d\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.564776 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-httpd-config\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.564845 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9chh\" (UniqueName: \"kubernetes.io/projected/0164c2f9-e32a-4c4e-b778-01f80e950d13-kube-api-access-b9chh\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.564911 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-config\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.565044 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-combined-ca-bundle\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.565086 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-ovndb-tls-certs\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.566469 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.669967 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-httpd-config\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.670032 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9chh\" (UniqueName: \"kubernetes.io/projected/0164c2f9-e32a-4c4e-b778-01f80e950d13-kube-api-access-b9chh\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.670085 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-config\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.670112 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-combined-ca-bundle\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.670130 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-ovndb-tls-certs\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.675805 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-httpd-config\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.688952 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-config\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.691409 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-ovndb-tls-certs\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.697266 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-combined-ca-bundle\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.698488 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9chh\" (UniqueName: \"kubernetes.io/projected/0164c2f9-e32a-4c4e-b778-01f80e950d13-kube-api-access-b9chh\") pod \"neutron-5c574bf78b-4thfs\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: E1217 08:39:36.708635 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-cinder-api:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:36 crc kubenswrapper[4966]: E1217 08:39:36.708696 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-cinder-api:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:39:36 crc kubenswrapper[4966]: E1217 08:39:36.708843 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-cinder-api:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9r6ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-rqvfb_openstack(3bca810d-b802-481a-8792-3be9b6a141bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:39:36 crc kubenswrapper[4966]: E1217 08:39:36.710021 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-rqvfb" podUID="3bca810d-b802-481a-8792-3be9b6a141bc" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.742338 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.764135 4966 scope.go:117] "RemoveContainer" containerID="9c14ed179cf5d0e21b47874bd77fdb37f9696a22277c7e74dc7b659e81a11798" Dec 17 08:39:36 crc kubenswrapper[4966]: W1217 08:39:36.772380 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04dbdf7d_ebdc_46b9_8b29_b6e427b5d017.slice/crio-7098111dbed79bb95427735d2299eb8c8e2ddb3e2ca55cf702267f4b20e66124 WatchSource:0}: Error finding container 7098111dbed79bb95427735d2299eb8c8e2ddb3e2ca55cf702267f4b20e66124: Status 404 returned error can't find the container with id 7098111dbed79bb95427735d2299eb8c8e2ddb3e2ca55cf702267f4b20e66124 Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.914557 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c007a81-a0c1-43d8-8a77-62aebd2f188d" path="/var/lib/kubelet/pods/0c007a81-a0c1-43d8-8a77-62aebd2f188d/volumes" Dec 17 08:39:36 crc kubenswrapper[4966]: I1217 08:39:36.917183 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" path="/var/lib/kubelet/pods/28af8586-3f33-42a1-813b-a1382cde5ec9/volumes" Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.053047 4966 scope.go:117] "RemoveContainer" containerID="dcc0ca43aaebd497499db3c8dcaaf4d3e08926ba98c3fe2efc92140eefde3ae2" Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.165325 4966 scope.go:117] "RemoveContainer" containerID="3d75e4eaace730cdfca14100800b1f7cd7e07756b219c4f27803144c093bc3e1" Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.294208 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55d57cbbcb-2tmk5"] Dec 17 08:39:37 crc kubenswrapper[4966]: W1217 08:39:37.373196 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3ade9e0_df6d_4b67_a0a3_727dc696a358.slice/crio-2a017328e901bb34df6a57c65656f3a06ffd485cd74840741e656bb523de307e WatchSource:0}: Error finding container 2a017328e901bb34df6a57c65656f3a06ffd485cd74840741e656bb523de307e: Status 404 returned error can't find the container with id 2a017328e901bb34df6a57c65656f3a06ffd485cd74840741e656bb523de307e Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.441198 4966 scope.go:117] "RemoveContainer" containerID="f54a4f58f7bb67f9d87e7addb7489a5b71d139a3f9c89fc2c158fe0bbd974742" Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.514078 4966 scope.go:117] "RemoveContainer" containerID="c135e436514d63ea87d326e287784898342c3f76fa4248d93984bc19d0699d82" Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.523833 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb7cd8c7-7q85d" event={"ID":"03c06b3f-4fda-455c-9101-ee90d055ed88","Type":"ContainerStarted","Data":"ff7e3d8aade2d5e0145c66ead1c98008d23c86c3029856c0633fea20e05670ea"} Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.540079 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerStarted","Data":"5a3ca13748e8803e6e39f736c666c4d4e73383bc2acc5ebdc79d6f5a01c7c2fd"} Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.540962 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fbfdd886-69frk" event={"ID":"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017","Type":"ContainerStarted","Data":"7098111dbed79bb95427735d2299eb8c8e2ddb3e2ca55cf702267f4b20e66124"} Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.542891 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55d57cbbcb-2tmk5" event={"ID":"f3ade9e0-df6d-4b67-a0a3-727dc696a358","Type":"ContainerStarted","Data":"2a017328e901bb34df6a57c65656f3a06ffd485cd74840741e656bb523de307e"} Dec 17 08:39:37 crc kubenswrapper[4966]: E1217 08:39:37.543407 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.156:5001/podified-antelope-centos9/openstack-cinder-api:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/cinder-db-sync-rqvfb" podUID="3bca810d-b802-481a-8792-3be9b6a141bc" Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.557278 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.786559 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gkjk7"] Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.798265 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f4484465c-9cs9d"] Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.807333 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:39:37 crc kubenswrapper[4966]: W1217 08:39:37.817985 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95497f6e_9ff6_457b_a47a_3747cc4ddc69.slice/crio-6ce509ba09b23ace646f206ae81ad457b3ba9fc24da29fae383f59b9c2a9bb8e WatchSource:0}: Error finding container 6ce509ba09b23ace646f206ae81ad457b3ba9fc24da29fae383f59b9c2a9bb8e: Status 404 returned error can't find the container with id 6ce509ba09b23ace646f206ae81ad457b3ba9fc24da29fae383f59b9c2a9bb8e Dec 17 08:39:37 crc kubenswrapper[4966]: I1217 08:39:37.923495 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c574bf78b-4thfs"] Dec 17 08:39:37 crc kubenswrapper[4966]: W1217 08:39:37.932041 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0164c2f9_e32a_4c4e_b778_01f80e950d13.slice/crio-befbc11302866673ff657cd5909e32efacf241a68c17a65985234752e2e1b319 WatchSource:0}: Error finding container befbc11302866673ff657cd5909e32efacf241a68c17a65985234752e2e1b319: Status 404 returned error can't find the container with id befbc11302866673ff657cd5909e32efacf241a68c17a65985234752e2e1b319 Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.573724 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6jj7z" event={"ID":"eccb295b-14f1-43a9-98c8-673e2ce78add","Type":"ContainerStarted","Data":"d6333c355cf08791bb1a8ac20b4ad7952b8619bcee1c6c44408c3f020e7e0342"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.578395 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"073e8454-1a1f-4a48-9516-070e0fddc046","Type":"ContainerStarted","Data":"75ab3382d75a5398870a29868b8a89564dfbb4e02a9cd6e837f5ea34d310a8c0"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.594563 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6jj7z" podStartSLOduration=4.081641375 podStartE2EDuration="41.594547312s" podCreationTimestamp="2025-12-17 08:38:57 +0000 UTC" firstStartedPulling="2025-12-17 08:38:59.486652177 +0000 UTC m=+1075.031722119" lastFinishedPulling="2025-12-17 08:39:36.999558104 +0000 UTC m=+1112.544628056" observedRunningTime="2025-12-17 08:39:38.592511377 +0000 UTC m=+1114.137581319" watchObservedRunningTime="2025-12-17 08:39:38.594547312 +0000 UTC m=+1114.139617254" Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.597111 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gkjk7" event={"ID":"95497f6e-9ff6-457b-a47a-3747cc4ddc69","Type":"ContainerStarted","Data":"6ce509ba09b23ace646f206ae81ad457b3ba9fc24da29fae383f59b9c2a9bb8e"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.605276 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f89dfb859-v8j22" event={"ID":"129625e4-4eef-46f8-861b-831451355683","Type":"ContainerStarted","Data":"9c00ba7277e221aa4226c948848d49a5be5d40c7e0b3871860ca5e45e38a152d"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.605319 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f89dfb859-v8j22" event={"ID":"129625e4-4eef-46f8-861b-831451355683","Type":"ContainerStarted","Data":"e4f442511505863b3d83e05feee7f4a5e4126aff413d9b1f0847ac54a6bcb393"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.605385 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f89dfb859-v8j22" podUID="129625e4-4eef-46f8-861b-831451355683" containerName="horizon-log" containerID="cri-o://e4f442511505863b3d83e05feee7f4a5e4126aff413d9b1f0847ac54a6bcb393" gracePeriod=30 Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.605563 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f89dfb859-v8j22" podUID="129625e4-4eef-46f8-861b-831451355683" containerName="horizon" containerID="cri-o://9c00ba7277e221aa4226c948848d49a5be5d40c7e0b3871860ca5e45e38a152d" gracePeriod=30 Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.614272 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fbfdd886-69frk" event={"ID":"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017","Type":"ContainerStarted","Data":"c24b7f60ad38b0edeedf9daa35fb224111274921103a9efaaa7ef438157c889c"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.632127 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55d57cbbcb-2tmk5" event={"ID":"f3ade9e0-df6d-4b67-a0a3-727dc696a358","Type":"ContainerStarted","Data":"bb1a767f1ab1b4b0cac568a1102b6ccd4213f8f91162f647356eb06b24453cfe"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.637323 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c574bf78b-4thfs" event={"ID":"0164c2f9-e32a-4c4e-b778-01f80e950d13","Type":"ContainerStarted","Data":"befbc11302866673ff657cd5909e32efacf241a68c17a65985234752e2e1b319"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.642820 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5f89dfb859-v8j22" podStartSLOduration=3.319077513 podStartE2EDuration="38.642802325s" podCreationTimestamp="2025-12-17 08:39:00 +0000 UTC" firstStartedPulling="2025-12-17 08:39:01.47767959 +0000 UTC m=+1077.022749532" lastFinishedPulling="2025-12-17 08:39:36.801404402 +0000 UTC m=+1112.346474344" observedRunningTime="2025-12-17 08:39:38.637298794 +0000 UTC m=+1114.182368746" watchObservedRunningTime="2025-12-17 08:39:38.642802325 +0000 UTC m=+1114.187872267" Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.647237 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7a7b745-08ce-4be8-b81b-82f632be09a3","Type":"ContainerStarted","Data":"8edf28b9737ddca025b20f80661e6bac0063e79b9f49e79027b45c0f9f760998"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.664079 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-fbfdd886-69frk" podStartSLOduration=32.664059418 podStartE2EDuration="32.664059418s" podCreationTimestamp="2025-12-17 08:39:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:38.663859422 +0000 UTC m=+1114.208929364" watchObservedRunningTime="2025-12-17 08:39:38.664059418 +0000 UTC m=+1114.209129360" Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.671543 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb7cd8c7-7q85d" event={"ID":"03c06b3f-4fda-455c-9101-ee90d055ed88","Type":"ContainerStarted","Data":"c816d2354e13f8ad6cb1d6ceb2a3704a74f632b96240b03f135b84d09ac5d48f"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.671721 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67cb7cd8c7-7q85d" podUID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerName="horizon-log" containerID="cri-o://ff7e3d8aade2d5e0145c66ead1c98008d23c86c3029856c0633fea20e05670ea" gracePeriod=30 Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.672097 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67cb7cd8c7-7q85d" podUID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerName="horizon" containerID="cri-o://c816d2354e13f8ad6cb1d6ceb2a3704a74f632b96240b03f135b84d09ac5d48f" gracePeriod=30 Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.677472 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" event={"ID":"cff2f826-ba4a-4fe1-9406-bfdb021aea19","Type":"ContainerStarted","Data":"6d49551a0c44807b31a589197dcae7aaa73441dc05ce38f34cf7eb08805bb58e"} Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.704001 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67cb7cd8c7-7q85d" podStartSLOduration=6.357176776 podStartE2EDuration="41.703984502s" podCreationTimestamp="2025-12-17 08:38:57 +0000 UTC" firstStartedPulling="2025-12-17 08:38:59.333347564 +0000 UTC m=+1074.878417506" lastFinishedPulling="2025-12-17 08:39:34.68015529 +0000 UTC m=+1110.225225232" observedRunningTime="2025-12-17 08:39:38.698048869 +0000 UTC m=+1114.243118811" watchObservedRunningTime="2025-12-17 08:39:38.703984502 +0000 UTC m=+1114.249054444" Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.910808 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7f575c747c-8xn4b"] Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.913080 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.916236 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.926131 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 17 08:39:38 crc kubenswrapper[4966]: I1217 08:39:38.943863 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f575c747c-8xn4b"] Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.068182 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-public-tls-certs\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.068542 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-httpd-config\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.068674 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-ovndb-tls-certs\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.068904 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-combined-ca-bundle\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.069042 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-config\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.069158 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-internal-tls-certs\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.069246 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htcd9\" (UniqueName: \"kubernetes.io/projected/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-kube-api-access-htcd9\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.170741 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-httpd-config\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.170795 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-ovndb-tls-certs\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.170824 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-combined-ca-bundle\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.170842 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-config\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.170880 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-internal-tls-certs\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.170902 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htcd9\" (UniqueName: \"kubernetes.io/projected/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-kube-api-access-htcd9\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.171234 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-public-tls-certs\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.185346 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-public-tls-certs\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.192944 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-internal-tls-certs\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.193634 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-config\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.195426 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-combined-ca-bundle\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.195601 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-ovndb-tls-certs\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.196156 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-httpd-config\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.201286 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htcd9\" (UniqueName: \"kubernetes.io/projected/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-kube-api-access-htcd9\") pod \"neutron-7f575c747c-8xn4b\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.283105 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.413361 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b75d776ff-f25pq" podUID="28af8586-3f33-42a1-813b-a1382cde5ec9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.687763 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fbfdd886-69frk" event={"ID":"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017","Type":"ContainerStarted","Data":"ff860a43024f6954192167e387c152dc2b8c1a2ad2cbeaee0c6229e8eb0dd637"} Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.691131 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55d57cbbcb-2tmk5" event={"ID":"f3ade9e0-df6d-4b67-a0a3-727dc696a358","Type":"ContainerStarted","Data":"997b21d97d06d4293b8bf357b64adda8bdf6a0cf44406057b5412354af5056e6"} Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.694499 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c574bf78b-4thfs" event={"ID":"0164c2f9-e32a-4c4e-b778-01f80e950d13","Type":"ContainerStarted","Data":"6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58"} Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.696434 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7a7b745-08ce-4be8-b81b-82f632be09a3","Type":"ContainerStarted","Data":"b0d8bf408d4225ebf88e7dd96f46b53365c4bc3cacf91efa323ad0b6e873dabd"} Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.698640 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"073e8454-1a1f-4a48-9516-070e0fddc046","Type":"ContainerStarted","Data":"14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119"} Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.700204 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gkjk7" event={"ID":"95497f6e-9ff6-457b-a47a-3747cc4ddc69","Type":"ContainerStarted","Data":"199905196c697327e46f3abf7b458fcc5d98f4c441c75c4866f5aadfc795f895"} Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.707722 4966 generic.go:334] "Generic (PLEG): container finished" podID="cff2f826-ba4a-4fe1-9406-bfdb021aea19" containerID="c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8" exitCode=0 Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.707769 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" event={"ID":"cff2f826-ba4a-4fe1-9406-bfdb021aea19","Type":"ContainerDied","Data":"c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8"} Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.718442 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-55d57cbbcb-2tmk5" podStartSLOduration=32.718412747 podStartE2EDuration="32.718412747s" podCreationTimestamp="2025-12-17 08:39:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:39.715338583 +0000 UTC m=+1115.260408525" watchObservedRunningTime="2025-12-17 08:39:39.718412747 +0000 UTC m=+1115.263482699" Dec 17 08:39:39 crc kubenswrapper[4966]: I1217 08:39:39.741322 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-gkjk7" podStartSLOduration=14.741304934 podStartE2EDuration="14.741304934s" podCreationTimestamp="2025-12-17 08:39:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:39.73421705 +0000 UTC m=+1115.279287012" watchObservedRunningTime="2025-12-17 08:39:39.741304934 +0000 UTC m=+1115.286374876" Dec 17 08:39:40 crc kubenswrapper[4966]: I1217 08:39:40.709579 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:39:42 crc kubenswrapper[4966]: I1217 08:39:42.631916 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f575c747c-8xn4b"] Dec 17 08:39:42 crc kubenswrapper[4966]: I1217 08:39:42.777494 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerStarted","Data":"7bddea5ae658bbc591e4d75963b5c75b508219d0c0ae56659eabe943b39f9767"} Dec 17 08:39:42 crc kubenswrapper[4966]: I1217 08:39:42.785542 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c574bf78b-4thfs" event={"ID":"0164c2f9-e32a-4c4e-b778-01f80e950d13","Type":"ContainerStarted","Data":"741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7"} Dec 17 08:39:42 crc kubenswrapper[4966]: I1217 08:39:42.786939 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:39:42 crc kubenswrapper[4966]: I1217 08:39:42.791751 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f575c747c-8xn4b" event={"ID":"bdefbb9f-3322-4e19-98bb-ab6c11b00eec","Type":"ContainerStarted","Data":"2f3055015c2555bf29cc516696fd3ed3eda444d57a36699d7aca76d64dde8dcf"} Dec 17 08:39:42 crc kubenswrapper[4966]: I1217 08:39:42.794258 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" event={"ID":"cff2f826-ba4a-4fe1-9406-bfdb021aea19","Type":"ContainerStarted","Data":"ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c"} Dec 17 08:39:42 crc kubenswrapper[4966]: I1217 08:39:42.795188 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:42 crc kubenswrapper[4966]: I1217 08:39:42.813890 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c574bf78b-4thfs" podStartSLOduration=6.8138580829999995 podStartE2EDuration="6.813858083s" podCreationTimestamp="2025-12-17 08:39:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:42.812513566 +0000 UTC m=+1118.357583508" watchObservedRunningTime="2025-12-17 08:39:42.813858083 +0000 UTC m=+1118.358928025" Dec 17 08:39:42 crc kubenswrapper[4966]: I1217 08:39:42.846565 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" podStartSLOduration=6.846540178 podStartE2EDuration="6.846540178s" podCreationTimestamp="2025-12-17 08:39:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:42.841659614 +0000 UTC m=+1118.386729576" watchObservedRunningTime="2025-12-17 08:39:42.846540178 +0000 UTC m=+1118.391610120" Dec 17 08:39:43 crc kubenswrapper[4966]: I1217 08:39:43.810383 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7a7b745-08ce-4be8-b81b-82f632be09a3","Type":"ContainerStarted","Data":"0dee864237b1ca3a178e7b9abb0c4951cf10120ef5333b8af086c605dd1a4ad6"} Dec 17 08:39:43 crc kubenswrapper[4966]: I1217 08:39:43.818384 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f575c747c-8xn4b" event={"ID":"bdefbb9f-3322-4e19-98bb-ab6c11b00eec","Type":"ContainerStarted","Data":"dcb9949ed79a7c30f72f4516d42dbe7ad4d4d56abffe4c13a496c110fc4c0627"} Dec 17 08:39:43 crc kubenswrapper[4966]: I1217 08:39:43.834552 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"073e8454-1a1f-4a48-9516-070e0fddc046","Type":"ContainerStarted","Data":"9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc"} Dec 17 08:39:43 crc kubenswrapper[4966]: I1217 08:39:43.846816 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=19.846794055 podStartE2EDuration="19.846794055s" podCreationTimestamp="2025-12-17 08:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:43.845908031 +0000 UTC m=+1119.390977973" watchObservedRunningTime="2025-12-17 08:39:43.846794055 +0000 UTC m=+1119.391863997" Dec 17 08:39:43 crc kubenswrapper[4966]: I1217 08:39:43.873690 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=19.873672272 podStartE2EDuration="19.873672272s" podCreationTimestamp="2025-12-17 08:39:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:43.871085161 +0000 UTC m=+1119.416155093" watchObservedRunningTime="2025-12-17 08:39:43.873672272 +0000 UTC m=+1119.418742214" Dec 17 08:39:44 crc kubenswrapper[4966]: I1217 08:39:44.863092 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f575c747c-8xn4b" event={"ID":"bdefbb9f-3322-4e19-98bb-ab6c11b00eec","Type":"ContainerStarted","Data":"58cc81c70881d77dbf0be09d4674cbc1ce6bb83b337e24017bece572a468bdfa"} Dec 17 08:39:44 crc kubenswrapper[4966]: I1217 08:39:44.864527 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:39:44 crc kubenswrapper[4966]: I1217 08:39:44.886222 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7f575c747c-8xn4b" podStartSLOduration=6.886204515 podStartE2EDuration="6.886204515s" podCreationTimestamp="2025-12-17 08:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:44.886056511 +0000 UTC m=+1120.431126473" watchObservedRunningTime="2025-12-17 08:39:44.886204515 +0000 UTC m=+1120.431274467" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.161691 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.161747 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.192382 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.192428 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.235809 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.246800 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.264641 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.282086 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.881642 4966 generic.go:334] "Generic (PLEG): container finished" podID="eccb295b-14f1-43a9-98c8-673e2ce78add" containerID="d6333c355cf08791bb1a8ac20b4ad7952b8619bcee1c6c44408c3f020e7e0342" exitCode=0 Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.881734 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6jj7z" event={"ID":"eccb295b-14f1-43a9-98c8-673e2ce78add","Type":"ContainerDied","Data":"d6333c355cf08791bb1a8ac20b4ad7952b8619bcee1c6c44408c3f020e7e0342"} Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.882191 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.882344 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.882359 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 17 08:39:45 crc kubenswrapper[4966]: I1217 08:39:45.882402 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 17 08:39:47 crc kubenswrapper[4966]: I1217 08:39:47.346122 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:47 crc kubenswrapper[4966]: I1217 08:39:47.346380 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:39:47 crc kubenswrapper[4966]: I1217 08:39:47.768226 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:47 crc kubenswrapper[4966]: I1217 08:39:47.769232 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:39:47 crc kubenswrapper[4966]: I1217 08:39:47.910432 4966 generic.go:334] "Generic (PLEG): container finished" podID="95497f6e-9ff6-457b-a47a-3747cc4ddc69" containerID="199905196c697327e46f3abf7b458fcc5d98f4c441c75c4866f5aadfc795f895" exitCode=0 Dec 17 08:39:47 crc kubenswrapper[4966]: I1217 08:39:47.910841 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gkjk7" event={"ID":"95497f6e-9ff6-457b-a47a-3747cc4ddc69","Type":"ContainerDied","Data":"199905196c697327e46f3abf7b458fcc5d98f4c441c75c4866f5aadfc795f895"} Dec 17 08:39:48 crc kubenswrapper[4966]: I1217 08:39:48.277188 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:39:50 crc kubenswrapper[4966]: I1217 08:39:50.507340 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 17 08:39:50 crc kubenswrapper[4966]: I1217 08:39:50.553646 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:51 crc kubenswrapper[4966]: I1217 08:39:51.596902 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:39:51 crc kubenswrapper[4966]: I1217 08:39:51.691465 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568fb5df6f-trz6g"] Dec 17 08:39:51 crc kubenswrapper[4966]: I1217 08:39:51.693184 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" podUID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" containerName="dnsmasq-dns" containerID="cri-o://6c8727b4b2f160093cb7b87a95b427dd52b19680f6d88b3dae14132e8c5bd476" gracePeriod=10 Dec 17 08:39:51 crc kubenswrapper[4966]: I1217 08:39:51.962612 4966 generic.go:334] "Generic (PLEG): container finished" podID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" containerID="6c8727b4b2f160093cb7b87a95b427dd52b19680f6d88b3dae14132e8c5bd476" exitCode=0 Dec 17 08:39:51 crc kubenswrapper[4966]: I1217 08:39:51.962656 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" event={"ID":"199ca5ed-4cae-4a6a-86c3-cf940a841b88","Type":"ContainerDied","Data":"6c8727b4b2f160093cb7b87a95b427dd52b19680f6d88b3dae14132e8c5bd476"} Dec 17 08:39:53 crc kubenswrapper[4966]: I1217 08:39:53.568090 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" podUID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.153:5353: connect: connection refused" Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.906077 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.916827 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6jj7z" Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.951788 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-scripts\") pod \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.951830 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eccb295b-14f1-43a9-98c8-673e2ce78add-logs\") pod \"eccb295b-14f1-43a9-98c8-673e2ce78add\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.951880 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp27h\" (UniqueName: \"kubernetes.io/projected/eccb295b-14f1-43a9-98c8-673e2ce78add-kube-api-access-lp27h\") pod \"eccb295b-14f1-43a9-98c8-673e2ce78add\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.951908 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-fernet-keys\") pod \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.951947 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-config-data\") pod \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.951965 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-credential-keys\") pod \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.951994 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-combined-ca-bundle\") pod \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.952020 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-combined-ca-bundle\") pod \"eccb295b-14f1-43a9-98c8-673e2ce78add\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.952045 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-config-data\") pod \"eccb295b-14f1-43a9-98c8-673e2ce78add\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.952065 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-scripts\") pod \"eccb295b-14f1-43a9-98c8-673e2ce78add\" (UID: \"eccb295b-14f1-43a9-98c8-673e2ce78add\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.952103 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb774\" (UniqueName: \"kubernetes.io/projected/95497f6e-9ff6-457b-a47a-3747cc4ddc69-kube-api-access-zb774\") pod \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\" (UID: \"95497f6e-9ff6-457b-a47a-3747cc4ddc69\") " Dec 17 08:39:54 crc kubenswrapper[4966]: I1217 08:39:54.958332 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eccb295b-14f1-43a9-98c8-673e2ce78add-logs" (OuterVolumeSpecName: "logs") pod "eccb295b-14f1-43a9-98c8-673e2ce78add" (UID: "eccb295b-14f1-43a9-98c8-673e2ce78add"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.035078 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-scripts" (OuterVolumeSpecName: "scripts") pod "eccb295b-14f1-43a9-98c8-673e2ce78add" (UID: "eccb295b-14f1-43a9-98c8-673e2ce78add"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.036295 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95497f6e-9ff6-457b-a47a-3747cc4ddc69-kube-api-access-zb774" (OuterVolumeSpecName: "kube-api-access-zb774") pod "95497f6e-9ff6-457b-a47a-3747cc4ddc69" (UID: "95497f6e-9ff6-457b-a47a-3747cc4ddc69"). InnerVolumeSpecName "kube-api-access-zb774". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.053122 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eccb295b-14f1-43a9-98c8-673e2ce78add-kube-api-access-lp27h" (OuterVolumeSpecName: "kube-api-access-lp27h") pod "eccb295b-14f1-43a9-98c8-673e2ce78add" (UID: "eccb295b-14f1-43a9-98c8-673e2ce78add"). InnerVolumeSpecName "kube-api-access-lp27h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.053386 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-scripts" (OuterVolumeSpecName: "scripts") pod "95497f6e-9ff6-457b-a47a-3747cc4ddc69" (UID: "95497f6e-9ff6-457b-a47a-3747cc4ddc69"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.053445 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "95497f6e-9ff6-457b-a47a-3747cc4ddc69" (UID: "95497f6e-9ff6-457b-a47a-3747cc4ddc69"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.053484 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "95497f6e-9ff6-457b-a47a-3747cc4ddc69" (UID: "95497f6e-9ff6-457b-a47a-3747cc4ddc69"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.056684 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.056708 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eccb295b-14f1-43a9-98c8-673e2ce78add-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.056716 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp27h\" (UniqueName: \"kubernetes.io/projected/eccb295b-14f1-43a9-98c8-673e2ce78add-kube-api-access-lp27h\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.056728 4966 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.056736 4966 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.056744 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.056753 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb774\" (UniqueName: \"kubernetes.io/projected/95497f6e-9ff6-457b-a47a-3747cc4ddc69-kube-api-access-zb774\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.066895 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-config-data" (OuterVolumeSpecName: "config-data") pod "95497f6e-9ff6-457b-a47a-3747cc4ddc69" (UID: "95497f6e-9ff6-457b-a47a-3747cc4ddc69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.069096 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gkjk7" event={"ID":"95497f6e-9ff6-457b-a47a-3747cc4ddc69","Type":"ContainerDied","Data":"6ce509ba09b23ace646f206ae81ad457b3ba9fc24da29fae383f59b9c2a9bb8e"} Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.069242 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ce509ba09b23ace646f206ae81ad457b3ba9fc24da29fae383f59b9c2a9bb8e" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.069444 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gkjk7" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.071235 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6jj7z" event={"ID":"eccb295b-14f1-43a9-98c8-673e2ce78add","Type":"ContainerDied","Data":"9313cce76f78eee5bc25b7c1276d1086b88b7bd7c22f478cc5aecb8519f655ed"} Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.071347 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9313cce76f78eee5bc25b7c1276d1086b88b7bd7c22f478cc5aecb8519f655ed" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.071348 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6jj7z" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.085038 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eccb295b-14f1-43a9-98c8-673e2ce78add" (UID: "eccb295b-14f1-43a9-98c8-673e2ce78add"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.108981 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.114396 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-config-data" (OuterVolumeSpecName: "config-data") pod "eccb295b-14f1-43a9-98c8-673e2ce78add" (UID: "eccb295b-14f1-43a9-98c8-673e2ce78add"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.114694 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95497f6e-9ff6-457b-a47a-3747cc4ddc69" (UID: "95497f6e-9ff6-457b-a47a-3747cc4ddc69"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.158297 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.158327 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95497f6e-9ff6-457b-a47a-3747cc4ddc69-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.158337 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.158346 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eccb295b-14f1-43a9-98c8-673e2ce78add-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.349065 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.469148 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-sb\") pod \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.469353 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-nb\") pod \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.469411 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-swift-storage-0\") pod \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.469455 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwbwh\" (UniqueName: \"kubernetes.io/projected/199ca5ed-4cae-4a6a-86c3-cf940a841b88-kube-api-access-gwbwh\") pod \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.469487 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-config\") pod \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.469768 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-svc\") pod \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\" (UID: \"199ca5ed-4cae-4a6a-86c3-cf940a841b88\") " Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.503620 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/199ca5ed-4cae-4a6a-86c3-cf940a841b88-kube-api-access-gwbwh" (OuterVolumeSpecName: "kube-api-access-gwbwh") pod "199ca5ed-4cae-4a6a-86c3-cf940a841b88" (UID: "199ca5ed-4cae-4a6a-86c3-cf940a841b88"). InnerVolumeSpecName "kube-api-access-gwbwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.576095 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwbwh\" (UniqueName: \"kubernetes.io/projected/199ca5ed-4cae-4a6a-86c3-cf940a841b88-kube-api-access-gwbwh\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.735472 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.736745 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "199ca5ed-4cae-4a6a-86c3-cf940a841b88" (UID: "199ca5ed-4cae-4a6a-86c3-cf940a841b88"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.786289 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.838903 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "199ca5ed-4cae-4a6a-86c3-cf940a841b88" (UID: "199ca5ed-4cae-4a6a-86c3-cf940a841b88"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.892743 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.913115 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "199ca5ed-4cae-4a6a-86c3-cf940a841b88" (UID: "199ca5ed-4cae-4a6a-86c3-cf940a841b88"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.937700 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "199ca5ed-4cae-4a6a-86c3-cf940a841b88" (UID: "199ca5ed-4cae-4a6a-86c3-cf940a841b88"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:55 crc kubenswrapper[4966]: I1217 08:39:55.992666 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-config" (OuterVolumeSpecName: "config") pod "199ca5ed-4cae-4a6a-86c3-cf940a841b88" (UID: "199ca5ed-4cae-4a6a-86c3-cf940a841b88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.001819 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.001906 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.001919 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/199ca5ed-4cae-4a6a-86c3-cf940a841b88-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.129584 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-879977d5b-rwf8b"] Dec 17 08:39:56 crc kubenswrapper[4966]: E1217 08:39:56.130080 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" containerName="dnsmasq-dns" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.130105 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" containerName="dnsmasq-dns" Dec 17 08:39:56 crc kubenswrapper[4966]: E1217 08:39:56.130123 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95497f6e-9ff6-457b-a47a-3747cc4ddc69" containerName="keystone-bootstrap" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.130130 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="95497f6e-9ff6-457b-a47a-3747cc4ddc69" containerName="keystone-bootstrap" Dec 17 08:39:56 crc kubenswrapper[4966]: E1217 08:39:56.130165 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" containerName="init" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.130171 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" containerName="init" Dec 17 08:39:56 crc kubenswrapper[4966]: E1217 08:39:56.130178 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eccb295b-14f1-43a9-98c8-673e2ce78add" containerName="placement-db-sync" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.130185 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="eccb295b-14f1-43a9-98c8-673e2ce78add" containerName="placement-db-sync" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.130349 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" containerName="dnsmasq-dns" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.130360 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="eccb295b-14f1-43a9-98c8-673e2ce78add" containerName="placement-db-sync" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.130373 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="95497f6e-9ff6-457b-a47a-3747cc4ddc69" containerName="keystone-bootstrap" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.161539 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-879977d5b-rwf8b"] Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.161645 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.175595 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.175916 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.175915 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.176154 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-czvn7" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.176262 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.176398 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.177997 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-595bcd9766-wkh6b"] Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.185949 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.191234 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" event={"ID":"199ca5ed-4cae-4a6a-86c3-cf940a841b88","Type":"ContainerDied","Data":"0ea705fe3f905a7f5c9c2ab549625001c1d12d75b5c98400ad978becf9d293cd"} Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.191304 4966 scope.go:117] "RemoveContainer" containerID="6c8727b4b2f160093cb7b87a95b427dd52b19680f6d88b3dae14132e8c5bd476" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.191459 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568fb5df6f-trz6g" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.216304 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-combined-ca-bundle\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.216349 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-public-tls-certs\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.216374 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-fernet-keys\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.216396 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-credential-keys\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.216423 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-config-data\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.216459 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk4tz\" (UniqueName: \"kubernetes.io/projected/ee6e729c-115c-4e76-8b1f-dea7cae4c698-kube-api-access-hk4tz\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.216493 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-scripts\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.216512 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-internal-tls-certs\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.219463 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.219698 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.219834 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.219962 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.220061 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sgmwm" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.242126 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9vf55" event={"ID":"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a","Type":"ContainerStarted","Data":"5c7c6d66f7d0164b94f4acbc81db1caba0184245b63f6bb8e83580d4f8db3546"} Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.263565 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-595bcd9766-wkh6b"] Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.266503 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9bd96" event={"ID":"b7351028-bc40-4197-baa1-9fb0c525170d","Type":"ContainerStarted","Data":"5babfdce6e01bd62acd946a7fa8de1a13d7e35a1722e62ce0e5bde6558519fef"} Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.288934 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerStarted","Data":"8dbc5e0821630f3a48a51d25a09ad40619167d96ffdaa221adf907da179071e2"} Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.321833 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-combined-ca-bundle\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.322668 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qrlk\" (UniqueName: \"kubernetes.io/projected/a255684b-468f-446b-a966-cabdbf9937bf-kube-api-access-2qrlk\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.322792 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk4tz\" (UniqueName: \"kubernetes.io/projected/ee6e729c-115c-4e76-8b1f-dea7cae4c698-kube-api-access-hk4tz\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.322944 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-scripts\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323119 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-internal-tls-certs\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323244 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a255684b-468f-446b-a966-cabdbf9937bf-logs\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323328 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-config-data\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323435 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-scripts\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323535 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-combined-ca-bundle\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323627 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-public-tls-certs\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323732 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-public-tls-certs\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323797 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-fernet-keys\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323882 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-internal-tls-certs\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.323964 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-credential-keys\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.324046 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-config-data\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.332852 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-fernet-keys\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.333141 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-credential-keys\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.322036 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568fb5df6f-trz6g"] Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.337415 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-combined-ca-bundle\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.338184 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-config-data\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.344547 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-internal-tls-certs\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.352023 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9vf55" podStartSLOduration=3.601532615 podStartE2EDuration="59.352006049s" podCreationTimestamp="2025-12-17 08:38:57 +0000 UTC" firstStartedPulling="2025-12-17 08:38:59.486949725 +0000 UTC m=+1075.032019667" lastFinishedPulling="2025-12-17 08:39:55.237423159 +0000 UTC m=+1130.782493101" observedRunningTime="2025-12-17 08:39:56.300326082 +0000 UTC m=+1131.845396024" watchObservedRunningTime="2025-12-17 08:39:56.352006049 +0000 UTC m=+1131.897076001" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.359898 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568fb5df6f-trz6g"] Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.362083 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-scripts\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.362280 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee6e729c-115c-4e76-8b1f-dea7cae4c698-public-tls-certs\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.369241 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk4tz\" (UniqueName: \"kubernetes.io/projected/ee6e729c-115c-4e76-8b1f-dea7cae4c698-kube-api-access-hk4tz\") pod \"keystone-879977d5b-rwf8b\" (UID: \"ee6e729c-115c-4e76-8b1f-dea7cae4c698\") " pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.396917 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-9bd96" podStartSLOduration=3.619344054 podStartE2EDuration="1m0.396898949s" podCreationTimestamp="2025-12-17 08:38:56 +0000 UTC" firstStartedPulling="2025-12-17 08:38:58.460039548 +0000 UTC m=+1074.005109490" lastFinishedPulling="2025-12-17 08:39:55.237594443 +0000 UTC m=+1130.782664385" observedRunningTime="2025-12-17 08:39:56.317173334 +0000 UTC m=+1131.862243276" watchObservedRunningTime="2025-12-17 08:39:56.396898949 +0000 UTC m=+1131.941968891" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.431194 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-config-data\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.431264 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-scripts\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.431318 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-public-tls-certs\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.431338 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-internal-tls-certs\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.431374 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-combined-ca-bundle\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.431390 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qrlk\" (UniqueName: \"kubernetes.io/projected/a255684b-468f-446b-a966-cabdbf9937bf-kube-api-access-2qrlk\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.431467 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a255684b-468f-446b-a966-cabdbf9937bf-logs\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.436127 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a255684b-468f-446b-a966-cabdbf9937bf-logs\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.447741 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-config-data\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.448432 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-internal-tls-certs\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.448486 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-scripts\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.464042 4966 scope.go:117] "RemoveContainer" containerID="6f0cdb24b2944687dac029575facceeb016def100367fbd904fe77f3c14e6ec1" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.464565 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-public-tls-certs\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.464928 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a255684b-468f-446b-a966-cabdbf9937bf-combined-ca-bundle\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.487343 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qrlk\" (UniqueName: \"kubernetes.io/projected/a255684b-468f-446b-a966-cabdbf9937bf-kube-api-access-2qrlk\") pod \"placement-595bcd9766-wkh6b\" (UID: \"a255684b-468f-446b-a966-cabdbf9937bf\") " pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.531256 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.562643 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:56 crc kubenswrapper[4966]: I1217 08:39:56.865616 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="199ca5ed-4cae-4a6a-86c3-cf940a841b88" path="/var/lib/kubelet/pods/199ca5ed-4cae-4a6a-86c3-cf940a841b88/volumes" Dec 17 08:39:57 crc kubenswrapper[4966]: I1217 08:39:57.205088 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-879977d5b-rwf8b"] Dec 17 08:39:57 crc kubenswrapper[4966]: I1217 08:39:57.214882 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-595bcd9766-wkh6b"] Dec 17 08:39:57 crc kubenswrapper[4966]: I1217 08:39:57.305728 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-595bcd9766-wkh6b" event={"ID":"a255684b-468f-446b-a966-cabdbf9937bf","Type":"ContainerStarted","Data":"c96c2d59e573bedf57096bc8f2425ef0284240cea82fd243c4ed87e153bfd9c5"} Dec 17 08:39:57 crc kubenswrapper[4966]: I1217 08:39:57.314322 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-879977d5b-rwf8b" event={"ID":"ee6e729c-115c-4e76-8b1f-dea7cae4c698","Type":"ContainerStarted","Data":"a4a9b36f0a3460c32857dfea21340b5da9e208a3cf46b6f052ea0585de228b80"} Dec 17 08:39:57 crc kubenswrapper[4966]: I1217 08:39:57.326914 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rqvfb" event={"ID":"3bca810d-b802-481a-8792-3be9b6a141bc","Type":"ContainerStarted","Data":"813da29eb9a6d3330fe50d3eef0a482abb8877d463e871b4ff9adaca45841ff3"} Dec 17 08:39:57 crc kubenswrapper[4966]: I1217 08:39:57.352215 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-fbfdd886-69frk" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.156:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.156:8443: connect: connection refused" Dec 17 08:39:57 crc kubenswrapper[4966]: I1217 08:39:57.362614 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-rqvfb" podStartSLOduration=3.915514602 podStartE2EDuration="1m0.362591529s" podCreationTimestamp="2025-12-17 08:38:57 +0000 UTC" firstStartedPulling="2025-12-17 08:38:58.778141557 +0000 UTC m=+1074.323211499" lastFinishedPulling="2025-12-17 08:39:55.225218484 +0000 UTC m=+1130.770288426" observedRunningTime="2025-12-17 08:39:57.349921711 +0000 UTC m=+1132.894991673" watchObservedRunningTime="2025-12-17 08:39:57.362591529 +0000 UTC m=+1132.907661471" Dec 17 08:39:57 crc kubenswrapper[4966]: I1217 08:39:57.782810 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-55d57cbbcb-2tmk5" podUID="f3ade9e0-df6d-4b67-a0a3-727dc696a358" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Dec 17 08:39:58 crc kubenswrapper[4966]: I1217 08:39:58.338256 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-879977d5b-rwf8b" event={"ID":"ee6e729c-115c-4e76-8b1f-dea7cae4c698","Type":"ContainerStarted","Data":"c2b9cad2db426b36a330866bffb1cec09e2c78c680ae8227b506ed3d11413412"} Dec 17 08:39:58 crc kubenswrapper[4966]: I1217 08:39:58.339296 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:39:58 crc kubenswrapper[4966]: I1217 08:39:58.343842 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-595bcd9766-wkh6b" event={"ID":"a255684b-468f-446b-a966-cabdbf9937bf","Type":"ContainerStarted","Data":"310b9ce097421857458d360fa3658232e05b22319b6cc9394c4c0422e2e0c149"} Dec 17 08:39:58 crc kubenswrapper[4966]: I1217 08:39:58.343939 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-595bcd9766-wkh6b" event={"ID":"a255684b-468f-446b-a966-cabdbf9937bf","Type":"ContainerStarted","Data":"7d47730a9976dcf6bcaac974a9e33deb5a4778e04ef3fed64ce302a80a548688"} Dec 17 08:39:58 crc kubenswrapper[4966]: I1217 08:39:58.344550 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:58 crc kubenswrapper[4966]: I1217 08:39:58.344579 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:39:58 crc kubenswrapper[4966]: I1217 08:39:58.359723 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-879977d5b-rwf8b" podStartSLOduration=2.359705779 podStartE2EDuration="2.359705779s" podCreationTimestamp="2025-12-17 08:39:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:58.359180004 +0000 UTC m=+1133.904249946" watchObservedRunningTime="2025-12-17 08:39:58.359705779 +0000 UTC m=+1133.904775721" Dec 17 08:39:58 crc kubenswrapper[4966]: I1217 08:39:58.380908 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-595bcd9766-wkh6b" podStartSLOduration=2.38088863 podStartE2EDuration="2.38088863s" podCreationTimestamp="2025-12-17 08:39:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:39:58.376479809 +0000 UTC m=+1133.921549761" watchObservedRunningTime="2025-12-17 08:39:58.38088863 +0000 UTC m=+1133.925958572" Dec 17 08:40:03 crc kubenswrapper[4966]: I1217 08:40:03.402813 4966 generic.go:334] "Generic (PLEG): container finished" podID="2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" containerID="5c7c6d66f7d0164b94f4acbc81db1caba0184245b63f6bb8e83580d4f8db3546" exitCode=0 Dec 17 08:40:03 crc kubenswrapper[4966]: I1217 08:40:03.402890 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9vf55" event={"ID":"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a","Type":"ContainerDied","Data":"5c7c6d66f7d0164b94f4acbc81db1caba0184245b63f6bb8e83580d4f8db3546"} Dec 17 08:40:05 crc kubenswrapper[4966]: I1217 08:40:05.461695 4966 generic.go:334] "Generic (PLEG): container finished" podID="b7351028-bc40-4197-baa1-9fb0c525170d" containerID="5babfdce6e01bd62acd946a7fa8de1a13d7e35a1722e62ce0e5bde6558519fef" exitCode=0 Dec 17 08:40:05 crc kubenswrapper[4966]: I1217 08:40:05.462090 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9bd96" event={"ID":"b7351028-bc40-4197-baa1-9fb0c525170d","Type":"ContainerDied","Data":"5babfdce6e01bd62acd946a7fa8de1a13d7e35a1722e62ce0e5bde6558519fef"} Dec 17 08:40:05 crc kubenswrapper[4966]: I1217 08:40:05.844265 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9vf55" Dec 17 08:40:05 crc kubenswrapper[4966]: I1217 08:40:05.941778 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-db-sync-config-data\") pod \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " Dec 17 08:40:05 crc kubenswrapper[4966]: I1217 08:40:05.942159 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-combined-ca-bundle\") pod \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " Dec 17 08:40:05 crc kubenswrapper[4966]: I1217 08:40:05.942232 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrr57\" (UniqueName: \"kubernetes.io/projected/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-kube-api-access-xrr57\") pod \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\" (UID: \"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a\") " Dec 17 08:40:05 crc kubenswrapper[4966]: I1217 08:40:05.954082 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" (UID: "2ceb72da-ea40-4c24-8b26-94b10ba9cc6a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:05 crc kubenswrapper[4966]: I1217 08:40:05.957117 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-kube-api-access-xrr57" (OuterVolumeSpecName: "kube-api-access-xrr57") pod "2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" (UID: "2ceb72da-ea40-4c24-8b26-94b10ba9cc6a"). InnerVolumeSpecName "kube-api-access-xrr57". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:06 crc kubenswrapper[4966]: I1217 08:40:06.046139 4966 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:06 crc kubenswrapper[4966]: I1217 08:40:06.046168 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrr57\" (UniqueName: \"kubernetes.io/projected/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-kube-api-access-xrr57\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:06 crc kubenswrapper[4966]: I1217 08:40:06.047828 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" (UID: "2ceb72da-ea40-4c24-8b26-94b10ba9cc6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:06 crc kubenswrapper[4966]: I1217 08:40:06.148976 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:06 crc kubenswrapper[4966]: I1217 08:40:06.472708 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9vf55" Dec 17 08:40:06 crc kubenswrapper[4966]: I1217 08:40:06.475050 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9vf55" event={"ID":"2ceb72da-ea40-4c24-8b26-94b10ba9cc6a","Type":"ContainerDied","Data":"69dacb6e019333d5f91c3ee578461bf033bacf167dbad0cbb83e6a155135d94b"} Dec 17 08:40:06 crc kubenswrapper[4966]: I1217 08:40:06.475113 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69dacb6e019333d5f91c3ee578461bf033bacf167dbad0cbb83e6a155135d94b" Dec 17 08:40:06 crc kubenswrapper[4966]: I1217 08:40:06.802851 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.039444 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-564669fcf5-gv2b4"] Dec 17 08:40:07 crc kubenswrapper[4966]: E1217 08:40:07.044195 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" containerName="barbican-db-sync" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.044228 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" containerName="barbican-db-sync" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.044525 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" containerName="barbican-db-sync" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.045533 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.050459 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.050657 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.050769 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-c7pcz" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.093987 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-564669fcf5-gv2b4"] Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.108535 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-d6b6d75b-q8xdx"] Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.110199 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.114408 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.157979 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d6b6d75b-q8xdx"] Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174432 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2gks\" (UniqueName: \"kubernetes.io/projected/752cdbef-e7f8-4835-9984-eeee31131963-kube-api-access-s2gks\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174496 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/752cdbef-e7f8-4835-9984-eeee31131963-logs\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174534 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5a721a1-6544-4175-9734-232023ca3f2f-config-data-custom\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174583 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq2t4\" (UniqueName: \"kubernetes.io/projected/f5a721a1-6544-4175-9734-232023ca3f2f-kube-api-access-xq2t4\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174612 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/752cdbef-e7f8-4835-9984-eeee31131963-config-data-custom\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174630 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/752cdbef-e7f8-4835-9984-eeee31131963-config-data\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174658 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5a721a1-6544-4175-9734-232023ca3f2f-combined-ca-bundle\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174676 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5a721a1-6544-4175-9734-232023ca3f2f-logs\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174693 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/752cdbef-e7f8-4835-9984-eeee31131963-combined-ca-bundle\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.174716 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5a721a1-6544-4175-9734-232023ca3f2f-config-data\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.229216 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5855cd8fb5-5jrnk"] Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.233793 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.257057 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5855cd8fb5-5jrnk"] Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.276428 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/752cdbef-e7f8-4835-9984-eeee31131963-config-data-custom\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.276465 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/752cdbef-e7f8-4835-9984-eeee31131963-config-data\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.276519 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5a721a1-6544-4175-9734-232023ca3f2f-combined-ca-bundle\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.276547 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5a721a1-6544-4175-9734-232023ca3f2f-logs\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.276574 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/752cdbef-e7f8-4835-9984-eeee31131963-combined-ca-bundle\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.276601 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5a721a1-6544-4175-9734-232023ca3f2f-config-data\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.276643 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2gks\" (UniqueName: \"kubernetes.io/projected/752cdbef-e7f8-4835-9984-eeee31131963-kube-api-access-s2gks\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.276692 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/752cdbef-e7f8-4835-9984-eeee31131963-logs\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.277148 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5a721a1-6544-4175-9734-232023ca3f2f-logs\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.278072 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5a721a1-6544-4175-9734-232023ca3f2f-config-data-custom\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.278140 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq2t4\" (UniqueName: \"kubernetes.io/projected/f5a721a1-6544-4175-9734-232023ca3f2f-kube-api-access-xq2t4\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.279664 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/752cdbef-e7f8-4835-9984-eeee31131963-logs\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.285178 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/752cdbef-e7f8-4835-9984-eeee31131963-combined-ca-bundle\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.291988 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5a721a1-6544-4175-9734-232023ca3f2f-config-data\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.299709 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/752cdbef-e7f8-4835-9984-eeee31131963-config-data-custom\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.301574 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5a721a1-6544-4175-9734-232023ca3f2f-config-data-custom\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.320567 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq2t4\" (UniqueName: \"kubernetes.io/projected/f5a721a1-6544-4175-9734-232023ca3f2f-kube-api-access-xq2t4\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.321271 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/752cdbef-e7f8-4835-9984-eeee31131963-config-data\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.323512 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5a721a1-6544-4175-9734-232023ca3f2f-combined-ca-bundle\") pod \"barbican-worker-564669fcf5-gv2b4\" (UID: \"f5a721a1-6544-4175-9734-232023ca3f2f\") " pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.358761 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-fbfdd886-69frk" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.156:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.156:8443: connect: connection refused" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.372553 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2gks\" (UniqueName: \"kubernetes.io/projected/752cdbef-e7f8-4835-9984-eeee31131963-kube-api-access-s2gks\") pod \"barbican-keystone-listener-d6b6d75b-q8xdx\" (UID: \"752cdbef-e7f8-4835-9984-eeee31131963\") " pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.378027 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5489d786c6-qz2lq"] Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.379603 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.379584 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.380103 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.380245 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-svc\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.380371 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-swift-storage-0\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.380427 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4hxb\" (UniqueName: \"kubernetes.io/projected/a47e6c6c-5917-4d21-b904-2d355e01c0f0-kube-api-access-n4hxb\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.380544 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-config\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.385725 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-564669fcf5-gv2b4" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.400721 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.422029 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5489d786c6-qz2lq"] Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.455346 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.481998 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.482041 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-combined-ca-bundle\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.482095 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.482124 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.482866 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.482967 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.483014 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df474304-225d-4dcc-8b31-05646232f0ae-logs\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.483078 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-svc\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.483688 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-svc\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.484823 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k872\" (UniqueName: \"kubernetes.io/projected/df474304-225d-4dcc-8b31-05646232f0ae-kube-api-access-6k872\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.484904 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-swift-storage-0\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.484941 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4hxb\" (UniqueName: \"kubernetes.io/projected/a47e6c6c-5917-4d21-b904-2d355e01c0f0-kube-api-access-n4hxb\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.485576 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data-custom\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.485635 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-config\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.485474 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-swift-storage-0\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.488818 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-config\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.514244 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4hxb\" (UniqueName: \"kubernetes.io/projected/a47e6c6c-5917-4d21-b904-2d355e01c0f0-kube-api-access-n4hxb\") pod \"dnsmasq-dns-5855cd8fb5-5jrnk\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.519624 4966 generic.go:334] "Generic (PLEG): container finished" podID="3bca810d-b802-481a-8792-3be9b6a141bc" containerID="813da29eb9a6d3330fe50d3eef0a482abb8877d463e871b4ff9adaca45841ff3" exitCode=0 Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.519705 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rqvfb" event={"ID":"3bca810d-b802-481a-8792-3be9b6a141bc","Type":"ContainerDied","Data":"813da29eb9a6d3330fe50d3eef0a482abb8877d463e871b4ff9adaca45841ff3"} Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.567121 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.591892 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-combined-ca-bundle\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.591985 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.592024 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df474304-225d-4dcc-8b31-05646232f0ae-logs\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.592237 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k872\" (UniqueName: \"kubernetes.io/projected/df474304-225d-4dcc-8b31-05646232f0ae-kube-api-access-6k872\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.592327 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data-custom\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.593072 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df474304-225d-4dcc-8b31-05646232f0ae-logs\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.604481 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.611711 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-combined-ca-bundle\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.631517 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data-custom\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.642254 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k872\" (UniqueName: \"kubernetes.io/projected/df474304-225d-4dcc-8b31-05646232f0ae-kube-api-access-6k872\") pod \"barbican-api-5489d786c6-qz2lq\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.748871 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:07 crc kubenswrapper[4966]: I1217 08:40:07.768298 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-55d57cbbcb-2tmk5" podUID="f3ade9e0-df6d-4b67-a0a3-727dc696a358" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Dec 17 08:40:08 crc kubenswrapper[4966]: I1217 08:40:08.532473 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-9bd96" event={"ID":"b7351028-bc40-4197-baa1-9fb0c525170d","Type":"ContainerDied","Data":"be7c2be87eeece0352b016ee9dd70da9ddb5be531807821350ec778950dc323e"} Dec 17 08:40:08 crc kubenswrapper[4966]: I1217 08:40:08.533483 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be7c2be87eeece0352b016ee9dd70da9ddb5be531807821350ec778950dc323e" Dec 17 08:40:08 crc kubenswrapper[4966]: I1217 08:40:08.573947 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9bd96" Dec 17 08:40:08 crc kubenswrapper[4966]: I1217 08:40:08.722134 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-config-data\") pod \"b7351028-bc40-4197-baa1-9fb0c525170d\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " Dec 17 08:40:08 crc kubenswrapper[4966]: I1217 08:40:08.722201 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-combined-ca-bundle\") pod \"b7351028-bc40-4197-baa1-9fb0c525170d\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " Dec 17 08:40:08 crc kubenswrapper[4966]: I1217 08:40:08.722710 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbq6s\" (UniqueName: \"kubernetes.io/projected/b7351028-bc40-4197-baa1-9fb0c525170d-kube-api-access-pbq6s\") pod \"b7351028-bc40-4197-baa1-9fb0c525170d\" (UID: \"b7351028-bc40-4197-baa1-9fb0c525170d\") " Dec 17 08:40:08 crc kubenswrapper[4966]: I1217 08:40:08.759208 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7351028-bc40-4197-baa1-9fb0c525170d-kube-api-access-pbq6s" (OuterVolumeSpecName: "kube-api-access-pbq6s") pod "b7351028-bc40-4197-baa1-9fb0c525170d" (UID: "b7351028-bc40-4197-baa1-9fb0c525170d"). InnerVolumeSpecName "kube-api-access-pbq6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:08 crc kubenswrapper[4966]: I1217 08:40:08.828069 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbq6s\" (UniqueName: \"kubernetes.io/projected/b7351028-bc40-4197-baa1-9fb0c525170d-kube-api-access-pbq6s\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.018485 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7351028-bc40-4197-baa1-9fb0c525170d" (UID: "b7351028-bc40-4197-baa1-9fb0c525170d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.032749 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.306265 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-config-data" (OuterVolumeSpecName: "config-data") pod "b7351028-bc40-4197-baa1-9fb0c525170d" (UID: "b7351028-bc40-4197-baa1-9fb0c525170d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.339908 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7351028-bc40-4197-baa1-9fb0c525170d-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.359556 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5855cd8fb5-5jrnk"] Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.359621 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.429353 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.441859 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c574bf78b-4thfs"] Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.442106 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c574bf78b-4thfs" podUID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerName="neutron-api" containerID="cri-o://6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58" gracePeriod=30 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.442246 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c574bf78b-4thfs" podUID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerName="neutron-httpd" containerID="cri-o://741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7" gracePeriod=30 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.542554 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-db-sync-config-data\") pod \"3bca810d-b802-481a-8792-3be9b6a141bc\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.542623 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-scripts\") pod \"3bca810d-b802-481a-8792-3be9b6a141bc\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.542662 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-combined-ca-bundle\") pod \"3bca810d-b802-481a-8792-3be9b6a141bc\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.542713 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bca810d-b802-481a-8792-3be9b6a141bc-etc-machine-id\") pod \"3bca810d-b802-481a-8792-3be9b6a141bc\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.542895 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r6ks\" (UniqueName: \"kubernetes.io/projected/3bca810d-b802-481a-8792-3be9b6a141bc-kube-api-access-9r6ks\") pod \"3bca810d-b802-481a-8792-3be9b6a141bc\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.542954 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-config-data\") pod \"3bca810d-b802-481a-8792-3be9b6a141bc\" (UID: \"3bca810d-b802-481a-8792-3be9b6a141bc\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.552981 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3bca810d-b802-481a-8792-3be9b6a141bc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3bca810d-b802-481a-8792-3be9b6a141bc" (UID: "3bca810d-b802-481a-8792-3be9b6a141bc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.558948 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3bca810d-b802-481a-8792-3be9b6a141bc" (UID: "3bca810d-b802-481a-8792-3be9b6a141bc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.569971 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-scripts" (OuterVolumeSpecName: "scripts") pod "3bca810d-b802-481a-8792-3be9b6a141bc" (UID: "3bca810d-b802-481a-8792-3be9b6a141bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.572373 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rqvfb" event={"ID":"3bca810d-b802-481a-8792-3be9b6a141bc","Type":"ContainerDied","Data":"18376422af54c8345710b38bd7d0b1937e1efee1168738c20edf7fe0a44981f1"} Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.572409 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18376422af54c8345710b38bd7d0b1937e1efee1168738c20edf7fe0a44981f1" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.572475 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rqvfb" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.607499 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bca810d-b802-481a-8792-3be9b6a141bc-kube-api-access-9r6ks" (OuterVolumeSpecName: "kube-api-access-9r6ks") pod "3bca810d-b802-481a-8792-3be9b6a141bc" (UID: "3bca810d-b802-481a-8792-3be9b6a141bc"). InnerVolumeSpecName "kube-api-access-9r6ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.607743 4966 generic.go:334] "Generic (PLEG): container finished" podID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerID="c816d2354e13f8ad6cb1d6ceb2a3704a74f632b96240b03f135b84d09ac5d48f" exitCode=137 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.607890 4966 generic.go:334] "Generic (PLEG): container finished" podID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerID="ff7e3d8aade2d5e0145c66ead1c98008d23c86c3029856c0633fea20e05670ea" exitCode=137 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.607933 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb7cd8c7-7q85d" event={"ID":"03c06b3f-4fda-455c-9101-ee90d055ed88","Type":"ContainerDied","Data":"c816d2354e13f8ad6cb1d6ceb2a3704a74f632b96240b03f135b84d09ac5d48f"} Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.607959 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb7cd8c7-7q85d" event={"ID":"03c06b3f-4fda-455c-9101-ee90d055ed88","Type":"ContainerDied","Data":"ff7e3d8aade2d5e0145c66ead1c98008d23c86c3029856c0633fea20e05670ea"} Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.628067 4966 generic.go:334] "Generic (PLEG): container finished" podID="129625e4-4eef-46f8-861b-831451355683" containerID="9c00ba7277e221aa4226c948848d49a5be5d40c7e0b3871860ca5e45e38a152d" exitCode=137 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.628102 4966 generic.go:334] "Generic (PLEG): container finished" podID="129625e4-4eef-46f8-861b-831451355683" containerID="e4f442511505863b3d83e05feee7f4a5e4126aff413d9b1f0847ac54a6bcb393" exitCode=137 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.628144 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f89dfb859-v8j22" event={"ID":"129625e4-4eef-46f8-861b-831451355683","Type":"ContainerDied","Data":"9c00ba7277e221aa4226c948848d49a5be5d40c7e0b3871860ca5e45e38a152d"} Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.628169 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f89dfb859-v8j22" event={"ID":"129625e4-4eef-46f8-861b-831451355683","Type":"ContainerDied","Data":"e4f442511505863b3d83e05feee7f4a5e4126aff413d9b1f0847ac54a6bcb393"} Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.646853 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r6ks\" (UniqueName: \"kubernetes.io/projected/3bca810d-b802-481a-8792-3be9b6a141bc-kube-api-access-9r6ks\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.652956 4966 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.653177 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.653254 4966 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bca810d-b802-481a-8792-3be9b6a141bc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.666935 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerStarted","Data":"5493e36eaf0a74e0ee898b72a7b70f252fb9beca3c6da8ed9c11248832a6fc69"} Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.667098 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="ceilometer-central-agent" containerID="cri-o://5a3ca13748e8803e6e39f736c666c4d4e73383bc2acc5ebdc79d6f5a01c7c2fd" gracePeriod=30 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.667325 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.667553 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="proxy-httpd" containerID="cri-o://5493e36eaf0a74e0ee898b72a7b70f252fb9beca3c6da8ed9c11248832a6fc69" gracePeriod=30 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.667596 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="sg-core" containerID="cri-o://8dbc5e0821630f3a48a51d25a09ad40619167d96ffdaa221adf907da179071e2" gracePeriod=30 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.667630 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="ceilometer-notification-agent" containerID="cri-o://7bddea5ae658bbc591e4d75963b5c75b508219d0c0ae56659eabe943b39f9767" gracePeriod=30 Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.696081 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bca810d-b802-481a-8792-3be9b6a141bc" (UID: "3bca810d-b802-481a-8792-3be9b6a141bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.702827 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-9bd96" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.704157 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-config-data" (OuterVolumeSpecName: "config-data") pod "3bca810d-b802-481a-8792-3be9b6a141bc" (UID: "3bca810d-b802-481a-8792-3be9b6a141bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.704323 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" event={"ID":"a47e6c6c-5917-4d21-b904-2d355e01c0f0","Type":"ContainerStarted","Data":"8793bc3e272a49578b5f47da1a5ed6568fbc70f63332ec27fab0e97c29ff2743"} Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.710832 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.180546446 podStartE2EDuration="1m12.71081291s" podCreationTimestamp="2025-12-17 08:38:57 +0000 UTC" firstStartedPulling="2025-12-17 08:38:59.343233695 +0000 UTC m=+1074.888303637" lastFinishedPulling="2025-12-17 08:40:08.873500159 +0000 UTC m=+1144.418570101" observedRunningTime="2025-12-17 08:40:09.69548062 +0000 UTC m=+1145.240550582" watchObservedRunningTime="2025-12-17 08:40:09.71081291 +0000 UTC m=+1145.255882852" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.716201 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.751239 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.754460 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.754500 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bca810d-b802-481a-8792-3be9b6a141bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857467 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqxv4\" (UniqueName: \"kubernetes.io/projected/129625e4-4eef-46f8-861b-831451355683-kube-api-access-bqxv4\") pod \"129625e4-4eef-46f8-861b-831451355683\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857566 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03c06b3f-4fda-455c-9101-ee90d055ed88-logs\") pod \"03c06b3f-4fda-455c-9101-ee90d055ed88\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857618 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/03c06b3f-4fda-455c-9101-ee90d055ed88-horizon-secret-key\") pod \"03c06b3f-4fda-455c-9101-ee90d055ed88\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857653 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-scripts\") pod \"129625e4-4eef-46f8-861b-831451355683\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857697 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129625e4-4eef-46f8-861b-831451355683-logs\") pod \"129625e4-4eef-46f8-861b-831451355683\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857722 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-scripts\") pod \"03c06b3f-4fda-455c-9101-ee90d055ed88\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857770 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-config-data\") pod \"03c06b3f-4fda-455c-9101-ee90d055ed88\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857810 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vkd2\" (UniqueName: \"kubernetes.io/projected/03c06b3f-4fda-455c-9101-ee90d055ed88-kube-api-access-8vkd2\") pod \"03c06b3f-4fda-455c-9101-ee90d055ed88\" (UID: \"03c06b3f-4fda-455c-9101-ee90d055ed88\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857837 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-config-data\") pod \"129625e4-4eef-46f8-861b-831451355683\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.857857 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/129625e4-4eef-46f8-861b-831451355683-horizon-secret-key\") pod \"129625e4-4eef-46f8-861b-831451355683\" (UID: \"129625e4-4eef-46f8-861b-831451355683\") " Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.872816 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/129625e4-4eef-46f8-861b-831451355683-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "129625e4-4eef-46f8-861b-831451355683" (UID: "129625e4-4eef-46f8-861b-831451355683"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.877155 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/129625e4-4eef-46f8-861b-831451355683-logs" (OuterVolumeSpecName: "logs") pod "129625e4-4eef-46f8-861b-831451355683" (UID: "129625e4-4eef-46f8-861b-831451355683"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.877324 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03c06b3f-4fda-455c-9101-ee90d055ed88-logs" (OuterVolumeSpecName: "logs") pod "03c06b3f-4fda-455c-9101-ee90d055ed88" (UID: "03c06b3f-4fda-455c-9101-ee90d055ed88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.894992 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03c06b3f-4fda-455c-9101-ee90d055ed88-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "03c06b3f-4fda-455c-9101-ee90d055ed88" (UID: "03c06b3f-4fda-455c-9101-ee90d055ed88"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.910021 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 17 08:40:09 crc kubenswrapper[4966]: E1217 08:40:09.910754 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bca810d-b802-481a-8792-3be9b6a141bc" containerName="cinder-db-sync" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.910814 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bca810d-b802-481a-8792-3be9b6a141bc" containerName="cinder-db-sync" Dec 17 08:40:09 crc kubenswrapper[4966]: E1217 08:40:09.910883 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerName="horizon-log" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.910933 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerName="horizon-log" Dec 17 08:40:09 crc kubenswrapper[4966]: E1217 08:40:09.910992 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="129625e4-4eef-46f8-861b-831451355683" containerName="horizon-log" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911047 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="129625e4-4eef-46f8-861b-831451355683" containerName="horizon-log" Dec 17 08:40:09 crc kubenswrapper[4966]: E1217 08:40:09.911110 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="129625e4-4eef-46f8-861b-831451355683" containerName="horizon" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911157 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="129625e4-4eef-46f8-861b-831451355683" containerName="horizon" Dec 17 08:40:09 crc kubenswrapper[4966]: E1217 08:40:09.911215 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerName="horizon" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911266 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerName="horizon" Dec 17 08:40:09 crc kubenswrapper[4966]: E1217 08:40:09.911320 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7351028-bc40-4197-baa1-9fb0c525170d" containerName="heat-db-sync" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911365 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7351028-bc40-4197-baa1-9fb0c525170d" containerName="heat-db-sync" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911598 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bca810d-b802-481a-8792-3be9b6a141bc" containerName="cinder-db-sync" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911662 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerName="horizon-log" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911716 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="129625e4-4eef-46f8-861b-831451355683" containerName="horizon-log" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911771 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7351028-bc40-4197-baa1-9fb0c525170d" containerName="heat-db-sync" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911824 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="129625e4-4eef-46f8-861b-831451355683" containerName="horizon" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.911902 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="03c06b3f-4fda-455c-9101-ee90d055ed88" containerName="horizon" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.912805 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.920483 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.920808 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.921001 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.929193 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kg8zx" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.967854 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p55qs\" (UniqueName: \"kubernetes.io/projected/ec394811-ff3a-49d5-86f7-6b736748f0c3-kube-api-access-p55qs\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.967911 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.967947 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec394811-ff3a-49d5-86f7-6b736748f0c3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.967982 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-scripts\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.968019 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.968221 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.968273 4966 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/03c06b3f-4fda-455c-9101-ee90d055ed88-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.968289 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129625e4-4eef-46f8-861b-831451355683-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.968297 4966 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/129625e4-4eef-46f8-861b-831451355683-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.968305 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03c06b3f-4fda-455c-9101-ee90d055ed88-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.993086 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/129625e4-4eef-46f8-861b-831451355683-kube-api-access-bqxv4" (OuterVolumeSpecName: "kube-api-access-bqxv4") pod "129625e4-4eef-46f8-861b-831451355683" (UID: "129625e4-4eef-46f8-861b-831451355683"). InnerVolumeSpecName "kube-api-access-bqxv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.996037 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03c06b3f-4fda-455c-9101-ee90d055ed88-kube-api-access-8vkd2" (OuterVolumeSpecName: "kube-api-access-8vkd2") pod "03c06b3f-4fda-455c-9101-ee90d055ed88" (UID: "03c06b3f-4fda-455c-9101-ee90d055ed88"). InnerVolumeSpecName "kube-api-access-8vkd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:09 crc kubenswrapper[4966]: I1217 08:40:09.997857 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5855cd8fb5-5jrnk"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.046949 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.069679 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.069962 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.070071 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p55qs\" (UniqueName: \"kubernetes.io/projected/ec394811-ff3a-49d5-86f7-6b736748f0c3-kube-api-access-p55qs\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.070155 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.070243 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec394811-ff3a-49d5-86f7-6b736748f0c3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.070328 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-scripts\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.070443 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqxv4\" (UniqueName: \"kubernetes.io/projected/129625e4-4eef-46f8-861b-831451355683-kube-api-access-bqxv4\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.070505 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vkd2\" (UniqueName: \"kubernetes.io/projected/03c06b3f-4fda-455c-9101-ee90d055ed88-kube-api-access-8vkd2\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.087788 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec394811-ff3a-49d5-86f7-6b736748f0c3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.111603 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.118222 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-scripts\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.138716 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.156590 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-scripts" (OuterVolumeSpecName: "scripts") pod "03c06b3f-4fda-455c-9101-ee90d055ed88" (UID: "03c06b3f-4fda-455c-9101-ee90d055ed88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.175062 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-798df684f5-9xzvn"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.176815 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.184001 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.227803 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d6b6d75b-q8xdx"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.227855 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-564669fcf5-gv2b4"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.227979 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.243242 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798df684f5-9xzvn"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.267492 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5489d786c6-qz2lq"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.279114 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8d9z\" (UniqueName: \"kubernetes.io/projected/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-kube-api-access-l8d9z\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.279154 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-sb\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.279179 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-nb\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.279269 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-swift-storage-0\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.279359 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-config\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.279399 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-svc\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.286820 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.292439 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-config-data" (OuterVolumeSpecName: "config-data") pod "03c06b3f-4fda-455c-9101-ee90d055ed88" (UID: "03c06b3f-4fda-455c-9101-ee90d055ed88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.293198 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p55qs\" (UniqueName: \"kubernetes.io/projected/ec394811-ff3a-49d5-86f7-6b736748f0c3-kube-api-access-p55qs\") pod \"cinder-scheduler-0\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.296405 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-config-data" (OuterVolumeSpecName: "config-data") pod "129625e4-4eef-46f8-861b-831451355683" (UID: "129625e4-4eef-46f8-861b-831451355683"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.305266 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.305333 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-scripts" (OuterVolumeSpecName: "scripts") pod "129625e4-4eef-46f8-861b-831451355683" (UID: "129625e4-4eef-46f8-861b-831451355683"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.307028 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.310047 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381096 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-config\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381352 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-svc\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381375 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8d9z\" (UniqueName: \"kubernetes.io/projected/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-kube-api-access-l8d9z\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381394 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-sb\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381410 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-nb\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381429 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-scripts\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381447 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q87pm\" (UniqueName: \"kubernetes.io/projected/58c789d1-d3cc-43f0-b9ee-27a0c4050663-kube-api-access-q87pm\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381470 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58c789d1-d3cc-43f0-b9ee-27a0c4050663-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381488 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c789d1-d3cc-43f0-b9ee-27a0c4050663-logs\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381593 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-swift-storage-0\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381661 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381701 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381724 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data-custom\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381766 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03c06b3f-4fda-455c-9101-ee90d055ed88-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381780 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.381791 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/129625e4-4eef-46f8-861b-831451355683-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.382244 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-config\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.386617 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-nb\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.386616 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-sb\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.387712 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-svc\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.387728 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-swift-storage-0\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.398719 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8d9z\" (UniqueName: \"kubernetes.io/projected/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-kube-api-access-l8d9z\") pod \"dnsmasq-dns-798df684f5-9xzvn\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.482943 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-scripts\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.482986 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q87pm\" (UniqueName: \"kubernetes.io/projected/58c789d1-d3cc-43f0-b9ee-27a0c4050663-kube-api-access-q87pm\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.483010 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58c789d1-d3cc-43f0-b9ee-27a0c4050663-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.483026 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c789d1-d3cc-43f0-b9ee-27a0c4050663-logs\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.483109 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.483142 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.483164 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data-custom\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.483197 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58c789d1-d3cc-43f0-b9ee-27a0c4050663-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.485476 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c789d1-d3cc-43f0-b9ee-27a0c4050663-logs\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.486556 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-scripts\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.499115 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.500696 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.500933 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data-custom\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.501106 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.505550 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q87pm\" (UniqueName: \"kubernetes.io/projected/58c789d1-d3cc-43f0-b9ee-27a0c4050663-kube-api-access-q87pm\") pod \"cinder-api-0\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.573922 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.619215 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.715283 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-589dbc6cbd-4hskt"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.716809 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.723144 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.723185 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.744581 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-589dbc6cbd-4hskt"] Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.758189 4966 generic.go:334] "Generic (PLEG): container finished" podID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerID="8dbc5e0821630f3a48a51d25a09ad40619167d96ffdaa221adf907da179071e2" exitCode=2 Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.758223 4966 generic.go:334] "Generic (PLEG): container finished" podID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerID="5a3ca13748e8803e6e39f736c666c4d4e73383bc2acc5ebdc79d6f5a01c7c2fd" exitCode=0 Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.758228 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerDied","Data":"8dbc5e0821630f3a48a51d25a09ad40619167d96ffdaa221adf907da179071e2"} Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.758277 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerDied","Data":"5a3ca13748e8803e6e39f736c666c4d4e73383bc2acc5ebdc79d6f5a01c7c2fd"} Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.767818 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" event={"ID":"752cdbef-e7f8-4835-9984-eeee31131963","Type":"ContainerStarted","Data":"88cd293b3e0af20c88da2d24a5f1a5c67f2547a89d7ec17f98d31748b724a7be"} Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.769100 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5489d786c6-qz2lq" event={"ID":"df474304-225d-4dcc-8b31-05646232f0ae","Type":"ContainerStarted","Data":"d48f5b81b22c3a5bbf1777ae8f52ecfb84e08dad6019d0925541ae512f59555e"} Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.787354 4966 generic.go:334] "Generic (PLEG): container finished" podID="a47e6c6c-5917-4d21-b904-2d355e01c0f0" containerID="6c56d5af004bcf1cc5024fc0d42d04f938bb8013634ec41ee70d605144518d74" exitCode=0 Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.787410 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" event={"ID":"a47e6c6c-5917-4d21-b904-2d355e01c0f0","Type":"ContainerDied","Data":"6c56d5af004bcf1cc5024fc0d42d04f938bb8013634ec41ee70d605144518d74"} Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.793719 4966 generic.go:334] "Generic (PLEG): container finished" podID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerID="741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7" exitCode=0 Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.793769 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c574bf78b-4thfs" event={"ID":"0164c2f9-e32a-4c4e-b778-01f80e950d13","Type":"ContainerDied","Data":"741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7"} Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.795563 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-564669fcf5-gv2b4" event={"ID":"f5a721a1-6544-4175-9734-232023ca3f2f","Type":"ContainerStarted","Data":"7c2c71e5fd38626e6806bc14dacf99e1d158f7b4f44796cc03f2a8d5d49129df"} Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.800862 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb7cd8c7-7q85d" event={"ID":"03c06b3f-4fda-455c-9101-ee90d055ed88","Type":"ContainerDied","Data":"68776d2dd2a1e6367cded822199becccc9a3601e85a8306064ab0456ec7f7be1"} Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.800932 4966 scope.go:117] "RemoveContainer" containerID="c816d2354e13f8ad6cb1d6ceb2a3704a74f632b96240b03f135b84d09ac5d48f" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.801088 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67cb7cd8c7-7q85d" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.808266 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfz7n\" (UniqueName: \"kubernetes.io/projected/b794efa0-cd44-46ef-bd39-5bcde98ec271-kube-api-access-xfz7n\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.808342 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-config-data-custom\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.808375 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-public-tls-certs\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.808410 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b794efa0-cd44-46ef-bd39-5bcde98ec271-logs\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.808437 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-internal-tls-certs\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.808526 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-config-data\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.808573 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-combined-ca-bundle\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.838161 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f89dfb859-v8j22" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.871843 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f89dfb859-v8j22" event={"ID":"129625e4-4eef-46f8-861b-831451355683","Type":"ContainerDied","Data":"2dd488b6a6dbfd7a5cb0294e6c995b6581844cbd8d4422168de836159d87af04"} Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.911171 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-public-tls-certs\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.911267 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b794efa0-cd44-46ef-bd39-5bcde98ec271-logs\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.911315 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-internal-tls-certs\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.911408 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-config-data\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.911435 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-combined-ca-bundle\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.911499 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfz7n\" (UniqueName: \"kubernetes.io/projected/b794efa0-cd44-46ef-bd39-5bcde98ec271-kube-api-access-xfz7n\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.911545 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-config-data-custom\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.913565 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b794efa0-cd44-46ef-bd39-5bcde98ec271-logs\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.920987 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-combined-ca-bundle\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.929329 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-config-data\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.933004 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-public-tls-certs\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.942574 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-internal-tls-certs\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.973716 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b794efa0-cd44-46ef-bd39-5bcde98ec271-config-data-custom\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.974301 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfz7n\" (UniqueName: \"kubernetes.io/projected/b794efa0-cd44-46ef-bd39-5bcde98ec271-kube-api-access-xfz7n\") pod \"barbican-api-589dbc6cbd-4hskt\" (UID: \"b794efa0-cd44-46ef-bd39-5bcde98ec271\") " pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:10 crc kubenswrapper[4966]: I1217 08:40:10.989855 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67cb7cd8c7-7q85d"] Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.037466 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67cb7cd8c7-7q85d"] Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.050721 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.075559 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f89dfb859-v8j22"] Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.104506 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f89dfb859-v8j22"] Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.191541 4966 scope.go:117] "RemoveContainer" containerID="ff7e3d8aade2d5e0145c66ead1c98008d23c86c3029856c0633fea20e05670ea" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.244527 4966 scope.go:117] "RemoveContainer" containerID="9c00ba7277e221aa4226c948848d49a5be5d40c7e0b3871860ca5e45e38a152d" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.424846 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.492012 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.529577 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-svc\") pod \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.530600 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-nb\") pod \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.530821 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4hxb\" (UniqueName: \"kubernetes.io/projected/a47e6c6c-5917-4d21-b904-2d355e01c0f0-kube-api-access-n4hxb\") pod \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.530956 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-config\") pod \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.531073 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-swift-storage-0\") pod \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.531138 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-sb\") pod \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\" (UID: \"a47e6c6c-5917-4d21-b904-2d355e01c0f0\") " Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.546808 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a47e6c6c-5917-4d21-b904-2d355e01c0f0-kube-api-access-n4hxb" (OuterVolumeSpecName: "kube-api-access-n4hxb") pod "a47e6c6c-5917-4d21-b904-2d355e01c0f0" (UID: "a47e6c6c-5917-4d21-b904-2d355e01c0f0"). InnerVolumeSpecName "kube-api-access-n4hxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.571711 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-config" (OuterVolumeSpecName: "config") pod "a47e6c6c-5917-4d21-b904-2d355e01c0f0" (UID: "a47e6c6c-5917-4d21-b904-2d355e01c0f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.583393 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a47e6c6c-5917-4d21-b904-2d355e01c0f0" (UID: "a47e6c6c-5917-4d21-b904-2d355e01c0f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.588751 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a47e6c6c-5917-4d21-b904-2d355e01c0f0" (UID: "a47e6c6c-5917-4d21-b904-2d355e01c0f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.622323 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a47e6c6c-5917-4d21-b904-2d355e01c0f0" (UID: "a47e6c6c-5917-4d21-b904-2d355e01c0f0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.623970 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a47e6c6c-5917-4d21-b904-2d355e01c0f0" (UID: "a47e6c6c-5917-4d21-b904-2d355e01c0f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.638478 4966 scope.go:117] "RemoveContainer" containerID="e4f442511505863b3d83e05feee7f4a5e4126aff413d9b1f0847ac54a6bcb393" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.639497 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.639524 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.639542 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.639551 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.639560 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a47e6c6c-5917-4d21-b904-2d355e01c0f0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.639568 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4hxb\" (UniqueName: \"kubernetes.io/projected/a47e6c6c-5917-4d21-b904-2d355e01c0f0-kube-api-access-n4hxb\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:11 crc kubenswrapper[4966]: W1217 08:40:11.673183 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec394811_ff3a_49d5_86f7_6b736748f0c3.slice/crio-98710729f88f076fd64bcab3bba6879e937917bd21027f055befc552e9feb050 WatchSource:0}: Error finding container 98710729f88f076fd64bcab3bba6879e937917bd21027f055befc552e9feb050: Status 404 returned error can't find the container with id 98710729f88f076fd64bcab3bba6879e937917bd21027f055befc552e9feb050 Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.777575 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798df684f5-9xzvn"] Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.799454 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.849597 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" event={"ID":"a47e6c6c-5917-4d21-b904-2d355e01c0f0","Type":"ContainerDied","Data":"8793bc3e272a49578b5f47da1a5ed6568fbc70f63332ec27fab0e97c29ff2743"} Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.849641 4966 scope.go:117] "RemoveContainer" containerID="6c56d5af004bcf1cc5024fc0d42d04f938bb8013634ec41ee70d605144518d74" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.849734 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5855cd8fb5-5jrnk" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.870767 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ec394811-ff3a-49d5-86f7-6b736748f0c3","Type":"ContainerStarted","Data":"98710729f88f076fd64bcab3bba6879e937917bd21027f055befc552e9feb050"} Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.899195 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5489d786c6-qz2lq" event={"ID":"df474304-225d-4dcc-8b31-05646232f0ae","Type":"ContainerStarted","Data":"5be1287005bdb64cb81c95fcde5bbc0a3fe0520e7547d2bd395fc41fd3351d74"} Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.899244 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5489d786c6-qz2lq" event={"ID":"df474304-225d-4dcc-8b31-05646232f0ae","Type":"ContainerStarted","Data":"441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5"} Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.899972 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:11 crc kubenswrapper[4966]: I1217 08:40:11.900072 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.019159 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5855cd8fb5-5jrnk"] Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.035643 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5855cd8fb5-5jrnk"] Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.077071 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-589dbc6cbd-4hskt"] Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.091685 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5489d786c6-qz2lq" podStartSLOduration=5.091661948 podStartE2EDuration="5.091661948s" podCreationTimestamp="2025-12-17 08:40:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:12.035517679 +0000 UTC m=+1147.580587621" watchObservedRunningTime="2025-12-17 08:40:12.091661948 +0000 UTC m=+1147.636731880" Dec 17 08:40:12 crc kubenswrapper[4966]: W1217 08:40:12.665428 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58c789d1_d3cc_43f0_b9ee_27a0c4050663.slice/crio-25949d611f7c47e72182867f762977e673454934a11f362b345707549cdd9b00 WatchSource:0}: Error finding container 25949d611f7c47e72182867f762977e673454934a11f362b345707549cdd9b00: Status 404 returned error can't find the container with id 25949d611f7c47e72182867f762977e673454934a11f362b345707549cdd9b00 Dec 17 08:40:12 crc kubenswrapper[4966]: W1217 08:40:12.675090 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3e02105_d411_4d80_9d37_ae44d4cd8a4c.slice/crio-0851c0256cf554c3732ecb7d2aa6d032f094cc221bb092bc3aec3214caf1e279 WatchSource:0}: Error finding container 0851c0256cf554c3732ecb7d2aa6d032f094cc221bb092bc3aec3214caf1e279: Status 404 returned error can't find the container with id 0851c0256cf554c3732ecb7d2aa6d032f094cc221bb092bc3aec3214caf1e279 Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.898334 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03c06b3f-4fda-455c-9101-ee90d055ed88" path="/var/lib/kubelet/pods/03c06b3f-4fda-455c-9101-ee90d055ed88/volumes" Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.900713 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="129625e4-4eef-46f8-861b-831451355683" path="/var/lib/kubelet/pods/129625e4-4eef-46f8-861b-831451355683/volumes" Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.901292 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a47e6c6c-5917-4d21-b904-2d355e01c0f0" path="/var/lib/kubelet/pods/a47e6c6c-5917-4d21-b904-2d355e01c0f0/volumes" Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.926947 4966 generic.go:334] "Generic (PLEG): container finished" podID="df474304-225d-4dcc-8b31-05646232f0ae" containerID="5be1287005bdb64cb81c95fcde5bbc0a3fe0520e7547d2bd395fc41fd3351d74" exitCode=1 Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.927024 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5489d786c6-qz2lq" event={"ID":"df474304-225d-4dcc-8b31-05646232f0ae","Type":"ContainerDied","Data":"5be1287005bdb64cb81c95fcde5bbc0a3fe0520e7547d2bd395fc41fd3351d74"} Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.927867 4966 scope.go:117] "RemoveContainer" containerID="5be1287005bdb64cb81c95fcde5bbc0a3fe0520e7547d2bd395fc41fd3351d74" Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.941260 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58c789d1-d3cc-43f0-b9ee-27a0c4050663","Type":"ContainerStarted","Data":"25949d611f7c47e72182867f762977e673454934a11f362b345707549cdd9b00"} Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.948971 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-589dbc6cbd-4hskt" event={"ID":"b794efa0-cd44-46ef-bd39-5bcde98ec271","Type":"ContainerStarted","Data":"d478fed9a8c7424590ece05ddfaedecc17ed7d5fa74f878a3632451639540d36"} Dec 17 08:40:12 crc kubenswrapper[4966]: I1217 08:40:12.962119 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" event={"ID":"e3e02105-d411-4d80-9d37-ae44d4cd8a4c","Type":"ContainerStarted","Data":"0851c0256cf554c3732ecb7d2aa6d032f094cc221bb092bc3aec3214caf1e279"} Dec 17 08:40:13 crc kubenswrapper[4966]: I1217 08:40:13.362155 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 17 08:40:13 crc kubenswrapper[4966]: I1217 08:40:13.749268 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:13 crc kubenswrapper[4966]: I1217 08:40:13.973168 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:40:13 crc kubenswrapper[4966]: I1217 08:40:13.982495 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-589dbc6cbd-4hskt" event={"ID":"b794efa0-cd44-46ef-bd39-5bcde98ec271","Type":"ContainerStarted","Data":"9a7251b4ccda278f44526dfde39d20b2607b46c7d87a4777073f3d717dde9daa"} Dec 17 08:40:13 crc kubenswrapper[4966]: I1217 08:40:13.984335 4966 generic.go:334] "Generic (PLEG): container finished" podID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerID="6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58" exitCode=0 Dec 17 08:40:13 crc kubenswrapper[4966]: I1217 08:40:13.984392 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c574bf78b-4thfs" event={"ID":"0164c2f9-e32a-4c4e-b778-01f80e950d13","Type":"ContainerDied","Data":"6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58"} Dec 17 08:40:13 crc kubenswrapper[4966]: I1217 08:40:13.984419 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c574bf78b-4thfs" event={"ID":"0164c2f9-e32a-4c4e-b778-01f80e950d13","Type":"ContainerDied","Data":"befbc11302866673ff657cd5909e32efacf241a68c17a65985234752e2e1b319"} Dec 17 08:40:13 crc kubenswrapper[4966]: I1217 08:40:13.984437 4966 scope.go:117] "RemoveContainer" containerID="741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7" Dec 17 08:40:13 crc kubenswrapper[4966]: I1217 08:40:13.984589 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c574bf78b-4thfs" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.023620 4966 scope.go:117] "RemoveContainer" containerID="6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.043034 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9chh\" (UniqueName: \"kubernetes.io/projected/0164c2f9-e32a-4c4e-b778-01f80e950d13-kube-api-access-b9chh\") pod \"0164c2f9-e32a-4c4e-b778-01f80e950d13\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.043195 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-ovndb-tls-certs\") pod \"0164c2f9-e32a-4c4e-b778-01f80e950d13\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.043698 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-combined-ca-bundle\") pod \"0164c2f9-e32a-4c4e-b778-01f80e950d13\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.043749 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-httpd-config\") pod \"0164c2f9-e32a-4c4e-b778-01f80e950d13\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.043825 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-config\") pod \"0164c2f9-e32a-4c4e-b778-01f80e950d13\" (UID: \"0164c2f9-e32a-4c4e-b778-01f80e950d13\") " Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.061939 4966 scope.go:117] "RemoveContainer" containerID="741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7" Dec 17 08:40:14 crc kubenswrapper[4966]: E1217 08:40:14.062353 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7\": container with ID starting with 741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7 not found: ID does not exist" containerID="741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.062392 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7"} err="failed to get container status \"741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7\": rpc error: code = NotFound desc = could not find container \"741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7\": container with ID starting with 741ba40ef4cd10d8e788fcb85b924061c1794cbd16c7b7788a3e00d4c89d35b7 not found: ID does not exist" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.062414 4966 scope.go:117] "RemoveContainer" containerID="6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58" Dec 17 08:40:14 crc kubenswrapper[4966]: E1217 08:40:14.062626 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58\": container with ID starting with 6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58 not found: ID does not exist" containerID="6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.062642 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58"} err="failed to get container status \"6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58\": rpc error: code = NotFound desc = could not find container \"6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58\": container with ID starting with 6aeab04ec4439b0f7439d05a065888f9ebd1f9b3675aad84d5196cb45bae0c58 not found: ID does not exist" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.072966 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0164c2f9-e32a-4c4e-b778-01f80e950d13" (UID: "0164c2f9-e32a-4c4e-b778-01f80e950d13"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.096204 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0164c2f9-e32a-4c4e-b778-01f80e950d13-kube-api-access-b9chh" (OuterVolumeSpecName: "kube-api-access-b9chh") pod "0164c2f9-e32a-4c4e-b778-01f80e950d13" (UID: "0164c2f9-e32a-4c4e-b778-01f80e950d13"). InnerVolumeSpecName "kube-api-access-b9chh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.147190 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9chh\" (UniqueName: \"kubernetes.io/projected/0164c2f9-e32a-4c4e-b778-01f80e950d13-kube-api-access-b9chh\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.147228 4966 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.285082 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-config" (OuterVolumeSpecName: "config") pod "0164c2f9-e32a-4c4e-b778-01f80e950d13" (UID: "0164c2f9-e32a-4c4e-b778-01f80e950d13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.300836 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0164c2f9-e32a-4c4e-b778-01f80e950d13" (UID: "0164c2f9-e32a-4c4e-b778-01f80e950d13"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.307859 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0164c2f9-e32a-4c4e-b778-01f80e950d13" (UID: "0164c2f9-e32a-4c4e-b778-01f80e950d13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.356744 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.357197 4966 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.357209 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0164c2f9-e32a-4c4e-b778-01f80e950d13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.628481 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c574bf78b-4thfs"] Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.639077 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5c574bf78b-4thfs"] Dec 17 08:40:14 crc kubenswrapper[4966]: I1217 08:40:14.882409 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0164c2f9-e32a-4c4e-b778-01f80e950d13" path="/var/lib/kubelet/pods/0164c2f9-e32a-4c4e-b778-01f80e950d13/volumes" Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.395839 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-589dbc6cbd-4hskt" event={"ID":"b794efa0-cd44-46ef-bd39-5bcde98ec271","Type":"ContainerStarted","Data":"5a94dce49887b1345af2e52adf2842d7c6c299d6e03a12084f730e9ddbcb591b"} Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.395890 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.396912 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.402593 4966 generic.go:334] "Generic (PLEG): container finished" podID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerID="6697217dbf113e688e6fd26cbe8283c10b8a4cf3beb83a634921c6f8a5a9916c" exitCode=0 Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.402648 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" event={"ID":"e3e02105-d411-4d80-9d37-ae44d4cd8a4c","Type":"ContainerDied","Data":"6697217dbf113e688e6fd26cbe8283c10b8a4cf3beb83a634921c6f8a5a9916c"} Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.413045 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ec394811-ff3a-49d5-86f7-6b736748f0c3","Type":"ContainerStarted","Data":"baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c"} Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.424446 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-589dbc6cbd-4hskt" podStartSLOduration=5.424431588 podStartE2EDuration="5.424431588s" podCreationTimestamp="2025-12-17 08:40:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:15.422191907 +0000 UTC m=+1150.967261849" watchObservedRunningTime="2025-12-17 08:40:15.424431588 +0000 UTC m=+1150.969501530" Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.455376 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-564669fcf5-gv2b4" event={"ID":"f5a721a1-6544-4175-9734-232023ca3f2f","Type":"ContainerStarted","Data":"078ae5ef652c5291dd94ac5c95abb566e04cb1221491ca393b855962ccf30bc5"} Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.455419 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-564669fcf5-gv2b4" event={"ID":"f5a721a1-6544-4175-9734-232023ca3f2f","Type":"ContainerStarted","Data":"3478663048b868591eed03c1afc524227a6e59751774c5cc1dd929add114681d"} Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.481125 4966 generic.go:334] "Generic (PLEG): container finished" podID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerID="7bddea5ae658bbc591e4d75963b5c75b508219d0c0ae56659eabe943b39f9767" exitCode=0 Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.481183 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerDied","Data":"7bddea5ae658bbc591e4d75963b5c75b508219d0c0ae56659eabe943b39f9767"} Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.487075 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-564669fcf5-gv2b4" podStartSLOduration=5.155794205 podStartE2EDuration="8.487051544s" podCreationTimestamp="2025-12-17 08:40:07 +0000 UTC" firstStartedPulling="2025-12-17 08:40:10.260609529 +0000 UTC m=+1145.805679471" lastFinishedPulling="2025-12-17 08:40:13.591866868 +0000 UTC m=+1149.136936810" observedRunningTime="2025-12-17 08:40:15.478908701 +0000 UTC m=+1151.023978643" watchObservedRunningTime="2025-12-17 08:40:15.487051544 +0000 UTC m=+1151.032121486" Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.495439 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" event={"ID":"752cdbef-e7f8-4835-9984-eeee31131963","Type":"ContainerStarted","Data":"15b0a88f90ee06de70ce9295601e7310a874739b73d2a2dcb7da90e080944359"} Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.495488 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" event={"ID":"752cdbef-e7f8-4835-9984-eeee31131963","Type":"ContainerStarted","Data":"b17c4d2b73cda5d87b2c88034644d1ea3b7f8ac80a441412f01cb6011ddaf691"} Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.508093 4966 generic.go:334] "Generic (PLEG): container finished" podID="df474304-225d-4dcc-8b31-05646232f0ae" containerID="e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864" exitCode=1 Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.508153 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5489d786c6-qz2lq" event={"ID":"df474304-225d-4dcc-8b31-05646232f0ae","Type":"ContainerDied","Data":"e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864"} Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.508183 4966 scope.go:117] "RemoveContainer" containerID="5be1287005bdb64cb81c95fcde5bbc0a3fe0520e7547d2bd395fc41fd3351d74" Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.508834 4966 scope.go:117] "RemoveContainer" containerID="e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864" Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.513745 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:15 crc kubenswrapper[4966]: E1217 08:40:15.517182 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-5489d786c6-qz2lq_openstack(df474304-225d-4dcc-8b31-05646232f0ae)\"" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.545740 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-d6b6d75b-q8xdx" podStartSLOduration=5.145675939 podStartE2EDuration="8.545724853s" podCreationTimestamp="2025-12-17 08:40:07 +0000 UTC" firstStartedPulling="2025-12-17 08:40:10.139408337 +0000 UTC m=+1145.684478279" lastFinishedPulling="2025-12-17 08:40:13.539457251 +0000 UTC m=+1149.084527193" observedRunningTime="2025-12-17 08:40:15.530495965 +0000 UTC m=+1151.075565897" watchObservedRunningTime="2025-12-17 08:40:15.545724853 +0000 UTC m=+1151.090794795" Dec 17 08:40:15 crc kubenswrapper[4966]: I1217 08:40:15.554621 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58c789d1-d3cc-43f0-b9ee-27a0c4050663","Type":"ContainerStarted","Data":"de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d"} Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.584782 4966 scope.go:117] "RemoveContainer" containerID="e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864" Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.584936 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:16 crc kubenswrapper[4966]: E1217 08:40:16.585321 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-5489d786c6-qz2lq_openstack(df474304-225d-4dcc-8b31-05646232f0ae)\"" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.587688 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58c789d1-d3cc-43f0-b9ee-27a0c4050663","Type":"ContainerStarted","Data":"1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42"} Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.587798 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerName="cinder-api-log" containerID="cri-o://de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d" gracePeriod=30 Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.587907 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.587949 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerName="cinder-api" containerID="cri-o://1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42" gracePeriod=30 Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.590938 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" event={"ID":"e3e02105-d411-4d80-9d37-ae44d4cd8a4c","Type":"ContainerStarted","Data":"37887270f964c260dd0744dae832d55f04a977e3e67d25a5dbe0c9863d3935dc"} Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.591094 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.596167 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ec394811-ff3a-49d5-86f7-6b736748f0c3","Type":"ContainerStarted","Data":"0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf"} Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.629305 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.629284213 podStartE2EDuration="6.629284213s" podCreationTimestamp="2025-12-17 08:40:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:16.627479694 +0000 UTC m=+1152.172549636" watchObservedRunningTime="2025-12-17 08:40:16.629284213 +0000 UTC m=+1152.174354155" Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.666418 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.828395362 podStartE2EDuration="7.66639792s" podCreationTimestamp="2025-12-17 08:40:09 +0000 UTC" firstStartedPulling="2025-12-17 08:40:11.680661763 +0000 UTC m=+1147.225731705" lastFinishedPulling="2025-12-17 08:40:13.518664321 +0000 UTC m=+1149.063734263" observedRunningTime="2025-12-17 08:40:16.648599742 +0000 UTC m=+1152.193669694" watchObservedRunningTime="2025-12-17 08:40:16.66639792 +0000 UTC m=+1152.211467862" Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.669331 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" podStartSLOduration=7.66931361 podStartE2EDuration="7.66931361s" podCreationTimestamp="2025-12-17 08:40:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:16.667816319 +0000 UTC m=+1152.212886261" watchObservedRunningTime="2025-12-17 08:40:16.66931361 +0000 UTC m=+1152.214383552" Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.749660 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:16 crc kubenswrapper[4966]: I1217 08:40:16.750172 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.183507 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.291538 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-scripts\") pod \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.291616 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data\") pod \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.291637 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data-custom\") pod \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.291691 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q87pm\" (UniqueName: \"kubernetes.io/projected/58c789d1-d3cc-43f0-b9ee-27a0c4050663-kube-api-access-q87pm\") pod \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.291712 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58c789d1-d3cc-43f0-b9ee-27a0c4050663-etc-machine-id\") pod \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.291749 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c789d1-d3cc-43f0-b9ee-27a0c4050663-logs\") pod \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.291802 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-combined-ca-bundle\") pod \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\" (UID: \"58c789d1-d3cc-43f0-b9ee-27a0c4050663\") " Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.292150 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58c789d1-d3cc-43f0-b9ee-27a0c4050663-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "58c789d1-d3cc-43f0-b9ee-27a0c4050663" (UID: "58c789d1-d3cc-43f0-b9ee-27a0c4050663"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.292923 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c789d1-d3cc-43f0-b9ee-27a0c4050663-logs" (OuterVolumeSpecName: "logs") pod "58c789d1-d3cc-43f0-b9ee-27a0c4050663" (UID: "58c789d1-d3cc-43f0-b9ee-27a0c4050663"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.293148 4966 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58c789d1-d3cc-43f0-b9ee-27a0c4050663-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.298674 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-scripts" (OuterVolumeSpecName: "scripts") pod "58c789d1-d3cc-43f0-b9ee-27a0c4050663" (UID: "58c789d1-d3cc-43f0-b9ee-27a0c4050663"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.300804 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c789d1-d3cc-43f0-b9ee-27a0c4050663-kube-api-access-q87pm" (OuterVolumeSpecName: "kube-api-access-q87pm") pod "58c789d1-d3cc-43f0-b9ee-27a0c4050663" (UID: "58c789d1-d3cc-43f0-b9ee-27a0c4050663"). InnerVolumeSpecName "kube-api-access-q87pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.301063 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "58c789d1-d3cc-43f0-b9ee-27a0c4050663" (UID: "58c789d1-d3cc-43f0-b9ee-27a0c4050663"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.323825 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58c789d1-d3cc-43f0-b9ee-27a0c4050663" (UID: "58c789d1-d3cc-43f0-b9ee-27a0c4050663"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.345280 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data" (OuterVolumeSpecName: "config-data") pod "58c789d1-d3cc-43f0-b9ee-27a0c4050663" (UID: "58c789d1-d3cc-43f0-b9ee-27a0c4050663"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.395546 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.395603 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.395617 4966 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.395635 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q87pm\" (UniqueName: \"kubernetes.io/projected/58c789d1-d3cc-43f0-b9ee-27a0c4050663-kube-api-access-q87pm\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.395647 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c789d1-d3cc-43f0-b9ee-27a0c4050663-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.395657 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c789d1-d3cc-43f0-b9ee-27a0c4050663-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.609845 4966 generic.go:334] "Generic (PLEG): container finished" podID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerID="1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42" exitCode=0 Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.609892 4966 generic.go:334] "Generic (PLEG): container finished" podID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerID="de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d" exitCode=143 Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.610015 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58c789d1-d3cc-43f0-b9ee-27a0c4050663","Type":"ContainerDied","Data":"1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42"} Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.610481 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58c789d1-d3cc-43f0-b9ee-27a0c4050663","Type":"ContainerDied","Data":"de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d"} Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.610523 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58c789d1-d3cc-43f0-b9ee-27a0c4050663","Type":"ContainerDied","Data":"25949d611f7c47e72182867f762977e673454934a11f362b345707549cdd9b00"} Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.610561 4966 scope.go:117] "RemoveContainer" containerID="1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.610800 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.611400 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.612214 4966 scope.go:117] "RemoveContainer" containerID="e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864" Dec 17 08:40:17 crc kubenswrapper[4966]: E1217 08:40:17.612640 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-5489d786c6-qz2lq_openstack(df474304-225d-4dcc-8b31-05646232f0ae)\"" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.666861 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.667703 4966 scope.go:117] "RemoveContainer" containerID="de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.682245 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.690813 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 17 08:40:17 crc kubenswrapper[4966]: E1217 08:40:17.708327 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerName="cinder-api-log" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708360 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerName="cinder-api-log" Dec 17 08:40:17 crc kubenswrapper[4966]: E1217 08:40:17.708384 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerName="neutron-httpd" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708392 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerName="neutron-httpd" Dec 17 08:40:17 crc kubenswrapper[4966]: E1217 08:40:17.708408 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a47e6c6c-5917-4d21-b904-2d355e01c0f0" containerName="init" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708417 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a47e6c6c-5917-4d21-b904-2d355e01c0f0" containerName="init" Dec 17 08:40:17 crc kubenswrapper[4966]: E1217 08:40:17.708443 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerName="neutron-api" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708450 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerName="neutron-api" Dec 17 08:40:17 crc kubenswrapper[4966]: E1217 08:40:17.708464 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerName="cinder-api" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708471 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerName="cinder-api" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708764 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerName="cinder-api-log" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708785 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerName="neutron-httpd" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708800 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0164c2f9-e32a-4c4e-b778-01f80e950d13" containerName="neutron-api" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708820 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a47e6c6c-5917-4d21-b904-2d355e01c0f0" containerName="init" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.708829 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" containerName="cinder-api" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.709948 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.710045 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.718351 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.718351 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.718674 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.725262 4966 scope.go:117] "RemoveContainer" containerID="1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42" Dec 17 08:40:17 crc kubenswrapper[4966]: E1217 08:40:17.725974 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42\": container with ID starting with 1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42 not found: ID does not exist" containerID="1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.726127 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42"} err="failed to get container status \"1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42\": rpc error: code = NotFound desc = could not find container \"1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42\": container with ID starting with 1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42 not found: ID does not exist" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.726229 4966 scope.go:117] "RemoveContainer" containerID="de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d" Dec 17 08:40:17 crc kubenswrapper[4966]: E1217 08:40:17.726664 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d\": container with ID starting with de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d not found: ID does not exist" containerID="de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.726693 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d"} err="failed to get container status \"de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d\": rpc error: code = NotFound desc = could not find container \"de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d\": container with ID starting with de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d not found: ID does not exist" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.726712 4966 scope.go:117] "RemoveContainer" containerID="1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.727052 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42"} err="failed to get container status \"1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42\": rpc error: code = NotFound desc = could not find container \"1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42\": container with ID starting with 1b5468e84b694d508d3fc2aff5187d57be6d2b6f3487470f28466726b9b1dd42 not found: ID does not exist" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.727077 4966 scope.go:117] "RemoveContainer" containerID="de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.727361 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d"} err="failed to get container status \"de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d\": rpc error: code = NotFound desc = could not find container \"de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d\": container with ID starting with de0edf034d95d4daab223301a4e475e7d17dfcf3bcfd68cae5f21e28790ec46d not found: ID does not exist" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.750530 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.757313 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.909915 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ac0b25e-d429-4ceb-aa8c-571112a863af-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.909967 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9v2v\" (UniqueName: \"kubernetes.io/projected/2ac0b25e-d429-4ceb-aa8c-571112a863af-kube-api-access-v9v2v\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.910031 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ac0b25e-d429-4ceb-aa8c-571112a863af-logs\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.910063 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-config-data\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.910093 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-scripts\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.910821 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.910946 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.910991 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:17 crc kubenswrapper[4966]: I1217 08:40:17.911027 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-config-data-custom\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012152 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ac0b25e-d429-4ceb-aa8c-571112a863af-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012223 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9v2v\" (UniqueName: \"kubernetes.io/projected/2ac0b25e-d429-4ceb-aa8c-571112a863af-kube-api-access-v9v2v\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012312 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ac0b25e-d429-4ceb-aa8c-571112a863af-logs\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012228 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ac0b25e-d429-4ceb-aa8c-571112a863af-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012354 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-config-data\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012398 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-scripts\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012439 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012479 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012511 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.012544 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-config-data-custom\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.013659 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ac0b25e-d429-4ceb-aa8c-571112a863af-logs\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.018317 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-config-data-custom\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.018901 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.020656 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-config-data\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.025122 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.025170 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.037367 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ac0b25e-d429-4ceb-aa8c-571112a863af-scripts\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.041921 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9v2v\" (UniqueName: \"kubernetes.io/projected/2ac0b25e-d429-4ceb-aa8c-571112a863af-kube-api-access-v9v2v\") pod \"cinder-api-0\" (UID: \"2ac0b25e-d429-4ceb-aa8c-571112a863af\") " pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.340389 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.629681 4966 scope.go:117] "RemoveContainer" containerID="e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.629778 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:18 crc kubenswrapper[4966]: E1217 08:40:18.630013 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-5489d786c6-qz2lq_openstack(df474304-225d-4dcc-8b31-05646232f0ae)\"" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.840562 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c789d1-d3cc-43f0-b9ee-27a0c4050663" path="/var/lib/kubelet/pods/58c789d1-d3cc-43f0-b9ee-27a0c4050663/volumes" Dec 17 08:40:18 crc kubenswrapper[4966]: I1217 08:40:18.888660 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 17 08:40:18 crc kubenswrapper[4966]: W1217 08:40:18.893856 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ac0b25e_d429_4ceb_aa8c_571112a863af.slice/crio-6b7fe40272732245a837663d02e023228943badd5a7e9071aab6d3b6d4a31a09 WatchSource:0}: Error finding container 6b7fe40272732245a837663d02e023228943badd5a7e9071aab6d3b6d4a31a09: Status 404 returned error can't find the container with id 6b7fe40272732245a837663d02e023228943badd5a7e9071aab6d3b6d4a31a09 Dec 17 08:40:19 crc kubenswrapper[4966]: I1217 08:40:19.358371 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:40:19 crc kubenswrapper[4966]: I1217 08:40:19.659419 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ac0b25e-d429-4ceb-aa8c-571112a863af","Type":"ContainerStarted","Data":"7ea3ca1374f29fa6cae3d1a54eee428720466dc360d118c0d7c7d75a5c76658b"} Dec 17 08:40:19 crc kubenswrapper[4966]: I1217 08:40:19.659801 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ac0b25e-d429-4ceb-aa8c-571112a863af","Type":"ContainerStarted","Data":"6b7fe40272732245a837663d02e023228943badd5a7e9071aab6d3b6d4a31a09"} Dec 17 08:40:19 crc kubenswrapper[4966]: I1217 08:40:19.751749 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:19 crc kubenswrapper[4966]: I1217 08:40:19.817356 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:40:20 crc kubenswrapper[4966]: I1217 08:40:20.499797 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 17 08:40:20 crc kubenswrapper[4966]: I1217 08:40:20.576078 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:40:20 crc kubenswrapper[4966]: I1217 08:40:20.640720 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f4484465c-9cs9d"] Dec 17 08:40:20 crc kubenswrapper[4966]: I1217 08:40:20.641046 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" podUID="cff2f826-ba4a-4fe1-9406-bfdb021aea19" containerName="dnsmasq-dns" containerID="cri-o://ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c" gracePeriod=10 Dec 17 08:40:20 crc kubenswrapper[4966]: I1217 08:40:20.748904 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ac0b25e-d429-4ceb-aa8c-571112a863af","Type":"ContainerStarted","Data":"bc7ae24e067fd0834f211d36d232e3f7fae88b8df1aa100ac563fb9e4aab96da"} Dec 17 08:40:20 crc kubenswrapper[4966]: I1217 08:40:20.750625 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 17 08:40:20 crc kubenswrapper[4966]: I1217 08:40:20.808492 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.808463553 podStartE2EDuration="3.808463553s" podCreationTimestamp="2025-12-17 08:40:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:20.785584665 +0000 UTC m=+1156.330654607" watchObservedRunningTime="2025-12-17 08:40:20.808463553 +0000 UTC m=+1156.353533495" Dec 17 08:40:20 crc kubenswrapper[4966]: I1217 08:40:20.951933 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:20.998821 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.319762 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.374557 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.481444 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-nb\") pod \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.481496 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-664fk\" (UniqueName: \"kubernetes.io/projected/cff2f826-ba4a-4fe1-9406-bfdb021aea19-kube-api-access-664fk\") pod \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.481691 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-config\") pod \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.481713 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-sb\") pod \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.481740 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-swift-storage-0\") pod \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.481760 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-svc\") pod \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.489079 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cff2f826-ba4a-4fe1-9406-bfdb021aea19-kube-api-access-664fk" (OuterVolumeSpecName: "kube-api-access-664fk") pod "cff2f826-ba4a-4fe1-9406-bfdb021aea19" (UID: "cff2f826-ba4a-4fe1-9406-bfdb021aea19"). InnerVolumeSpecName "kube-api-access-664fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.568606 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cff2f826-ba4a-4fe1-9406-bfdb021aea19" (UID: "cff2f826-ba4a-4fe1-9406-bfdb021aea19"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.582966 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cff2f826-ba4a-4fe1-9406-bfdb021aea19" (UID: "cff2f826-ba4a-4fe1-9406-bfdb021aea19"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.583696 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-nb\") pod \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\" (UID: \"cff2f826-ba4a-4fe1-9406-bfdb021aea19\") " Dec 17 08:40:21 crc kubenswrapper[4966]: W1217 08:40:21.583813 4966 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/cff2f826-ba4a-4fe1-9406-bfdb021aea19/volumes/kubernetes.io~configmap/ovsdbserver-nb Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.583827 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cff2f826-ba4a-4fe1-9406-bfdb021aea19" (UID: "cff2f826-ba4a-4fe1-9406-bfdb021aea19"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.584277 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.584296 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.584307 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-664fk\" (UniqueName: \"kubernetes.io/projected/cff2f826-ba4a-4fe1-9406-bfdb021aea19-kube-api-access-664fk\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.609297 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cff2f826-ba4a-4fe1-9406-bfdb021aea19" (UID: "cff2f826-ba4a-4fe1-9406-bfdb021aea19"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.622427 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cff2f826-ba4a-4fe1-9406-bfdb021aea19" (UID: "cff2f826-ba4a-4fe1-9406-bfdb021aea19"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.624266 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-config" (OuterVolumeSpecName: "config") pod "cff2f826-ba4a-4fe1-9406-bfdb021aea19" (UID: "cff2f826-ba4a-4fe1-9406-bfdb021aea19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.689854 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.689919 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.689929 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cff2f826-ba4a-4fe1-9406-bfdb021aea19-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.758272 4966 generic.go:334] "Generic (PLEG): container finished" podID="cff2f826-ba4a-4fe1-9406-bfdb021aea19" containerID="ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c" exitCode=0 Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.758776 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerName="cinder-scheduler" containerID="cri-o://baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c" gracePeriod=30 Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.758542 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.758434 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" event={"ID":"cff2f826-ba4a-4fe1-9406-bfdb021aea19","Type":"ContainerDied","Data":"ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c"} Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.759672 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f4484465c-9cs9d" event={"ID":"cff2f826-ba4a-4fe1-9406-bfdb021aea19","Type":"ContainerDied","Data":"6d49551a0c44807b31a589197dcae7aaa73441dc05ce38f34cf7eb08805bb58e"} Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.759745 4966 scope.go:117] "RemoveContainer" containerID="ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c" Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.761203 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerName="probe" containerID="cri-o://0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf" gracePeriod=30 Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.847539 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f4484465c-9cs9d"] Dec 17 08:40:21 crc kubenswrapper[4966]: I1217 08:40:21.863827 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f4484465c-9cs9d"] Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.142749 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-55d57cbbcb-2tmk5" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.189894 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-fbfdd886-69frk"] Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.190103 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-fbfdd886-69frk" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon-log" containerID="cri-o://c24b7f60ad38b0edeedf9daa35fb224111274921103a9efaaa7ef438157c889c" gracePeriod=30 Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.190495 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-fbfdd886-69frk" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon" containerID="cri-o://ff860a43024f6954192167e387c152dc2b8c1a2ad2cbeaee0c6229e8eb0dd637" gracePeriod=30 Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.403469 4966 scope.go:117] "RemoveContainer" containerID="c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.493714 4966 scope.go:117] "RemoveContainer" containerID="ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c" Dec 17 08:40:22 crc kubenswrapper[4966]: E1217 08:40:22.494481 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c\": container with ID starting with ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c not found: ID does not exist" containerID="ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.494517 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c"} err="failed to get container status \"ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c\": rpc error: code = NotFound desc = could not find container \"ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c\": container with ID starting with ec70f5dc4b5deabdb075d3ae8db5d65aa887a1283cdb6341179ab3702451348c not found: ID does not exist" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.494544 4966 scope.go:117] "RemoveContainer" containerID="c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8" Dec 17 08:40:22 crc kubenswrapper[4966]: E1217 08:40:22.494825 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8\": container with ID starting with c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8 not found: ID does not exist" containerID="c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.494853 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8"} err="failed to get container status \"c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8\": rpc error: code = NotFound desc = could not find container \"c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8\": container with ID starting with c98d1fa99350d3c3c8f90646faa532a1e301e7d01ac9336451c4acc03566a4f8 not found: ID does not exist" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.749821 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.749911 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.750492 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="barbican-api-log" containerStatusID={"Type":"cri-o","ID":"441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5"} pod="openstack/barbican-api-5489d786c6-qz2lq" containerMessage="Container barbican-api-log failed liveness probe, will be restarted" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.750516 4966 scope.go:117] "RemoveContainer" containerID="e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.750536 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" containerID="cri-o://441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5" gracePeriod=30 Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.750845 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.751049 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.169:9311/healthcheck\": dial tcp 10.217.0.169:9311: connect: connection refused" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.774350 4966 generic.go:334] "Generic (PLEG): container finished" podID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerID="0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf" exitCode=0 Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.774466 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ec394811-ff3a-49d5-86f7-6b736748f0c3","Type":"ContainerDied","Data":"0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf"} Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.791825 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-589dbc6cbd-4hskt" podUID="b794efa0-cd44-46ef-bd39-5bcde98ec271" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 08:40:22 crc kubenswrapper[4966]: I1217 08:40:22.842009 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cff2f826-ba4a-4fe1-9406-bfdb021aea19" path="/var/lib/kubelet/pods/cff2f826-ba4a-4fe1-9406-bfdb021aea19/volumes" Dec 17 08:40:23 crc kubenswrapper[4966]: E1217 08:40:23.185474 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-5489d786c6-qz2lq_openstack(df474304-225d-4dcc-8b31-05646232f0ae)\"" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.414644 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.540986 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec394811-ff3a-49d5-86f7-6b736748f0c3-etc-machine-id\") pod \"ec394811-ff3a-49d5-86f7-6b736748f0c3\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.541096 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data\") pod \"ec394811-ff3a-49d5-86f7-6b736748f0c3\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.541121 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-combined-ca-bundle\") pod \"ec394811-ff3a-49d5-86f7-6b736748f0c3\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.541213 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-scripts\") pod \"ec394811-ff3a-49d5-86f7-6b736748f0c3\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.541282 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data-custom\") pod \"ec394811-ff3a-49d5-86f7-6b736748f0c3\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.541355 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p55qs\" (UniqueName: \"kubernetes.io/projected/ec394811-ff3a-49d5-86f7-6b736748f0c3-kube-api-access-p55qs\") pod \"ec394811-ff3a-49d5-86f7-6b736748f0c3\" (UID: \"ec394811-ff3a-49d5-86f7-6b736748f0c3\") " Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.547633 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ec394811-ff3a-49d5-86f7-6b736748f0c3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ec394811-ff3a-49d5-86f7-6b736748f0c3" (UID: "ec394811-ff3a-49d5-86f7-6b736748f0c3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.550813 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ec394811-ff3a-49d5-86f7-6b736748f0c3" (UID: "ec394811-ff3a-49d5-86f7-6b736748f0c3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.551570 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec394811-ff3a-49d5-86f7-6b736748f0c3-kube-api-access-p55qs" (OuterVolumeSpecName: "kube-api-access-p55qs") pod "ec394811-ff3a-49d5-86f7-6b736748f0c3" (UID: "ec394811-ff3a-49d5-86f7-6b736748f0c3"). InnerVolumeSpecName "kube-api-access-p55qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.552626 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-scripts" (OuterVolumeSpecName: "scripts") pod "ec394811-ff3a-49d5-86f7-6b736748f0c3" (UID: "ec394811-ff3a-49d5-86f7-6b736748f0c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.572138 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.604515 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec394811-ff3a-49d5-86f7-6b736748f0c3" (UID: "ec394811-ff3a-49d5-86f7-6b736748f0c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.643271 4966 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.643302 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p55qs\" (UniqueName: \"kubernetes.io/projected/ec394811-ff3a-49d5-86f7-6b736748f0c3-kube-api-access-p55qs\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.643314 4966 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ec394811-ff3a-49d5-86f7-6b736748f0c3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.643322 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.643334 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.658126 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data" (OuterVolumeSpecName: "config-data") pod "ec394811-ff3a-49d5-86f7-6b736748f0c3" (UID: "ec394811-ff3a-49d5-86f7-6b736748f0c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.665096 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-589dbc6cbd-4hskt" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.756132 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec394811-ff3a-49d5-86f7-6b736748f0c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.769071 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5489d786c6-qz2lq"] Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.810281 4966 generic.go:334] "Generic (PLEG): container finished" podID="df474304-225d-4dcc-8b31-05646232f0ae" containerID="441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5" exitCode=143 Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.811242 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5489d786c6-qz2lq" event={"ID":"df474304-225d-4dcc-8b31-05646232f0ae","Type":"ContainerDied","Data":"441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5"} Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.811273 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5489d786c6-qz2lq" event={"ID":"df474304-225d-4dcc-8b31-05646232f0ae","Type":"ContainerStarted","Data":"b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a"} Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.811418 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5489d786c6-qz2lq" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" containerID="cri-o://b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a" gracePeriod=30 Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.811504 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.846039 4966 generic.go:334] "Generic (PLEG): container finished" podID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerID="baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c" exitCode=0 Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.846840 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.856918 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ec394811-ff3a-49d5-86f7-6b736748f0c3","Type":"ContainerDied","Data":"baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c"} Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.856975 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ec394811-ff3a-49d5-86f7-6b736748f0c3","Type":"ContainerDied","Data":"98710729f88f076fd64bcab3bba6879e937917bd21027f055befc552e9feb050"} Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.856995 4966 scope.go:117] "RemoveContainer" containerID="0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.954840 4966 scope.go:117] "RemoveContainer" containerID="baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c" Dec 17 08:40:23 crc kubenswrapper[4966]: I1217 08:40:23.971613 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.007386 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.027908 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 17 08:40:24 crc kubenswrapper[4966]: E1217 08:40:24.028351 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerName="probe" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.028366 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerName="probe" Dec 17 08:40:24 crc kubenswrapper[4966]: E1217 08:40:24.028385 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cff2f826-ba4a-4fe1-9406-bfdb021aea19" containerName="init" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.028390 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="cff2f826-ba4a-4fe1-9406-bfdb021aea19" containerName="init" Dec 17 08:40:24 crc kubenswrapper[4966]: E1217 08:40:24.028403 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cff2f826-ba4a-4fe1-9406-bfdb021aea19" containerName="dnsmasq-dns" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.028409 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="cff2f826-ba4a-4fe1-9406-bfdb021aea19" containerName="dnsmasq-dns" Dec 17 08:40:24 crc kubenswrapper[4966]: E1217 08:40:24.028428 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerName="cinder-scheduler" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.028434 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerName="cinder-scheduler" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.028754 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="cff2f826-ba4a-4fe1-9406-bfdb021aea19" containerName="dnsmasq-dns" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.028764 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerName="cinder-scheduler" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.028781 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec394811-ff3a-49d5-86f7-6b736748f0c3" containerName="probe" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.033909 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.043677 4966 scope.go:117] "RemoveContainer" containerID="0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.044028 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 17 08:40:24 crc kubenswrapper[4966]: E1217 08:40:24.050395 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf\": container with ID starting with 0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf not found: ID does not exist" containerID="0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.051394 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf"} err="failed to get container status \"0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf\": rpc error: code = NotFound desc = could not find container \"0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf\": container with ID starting with 0442cbb89b4a059ea20e3b9a56c76c8da840dbcb18e1b2da498e87ecebfaf9bf not found: ID does not exist" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.051498 4966 scope.go:117] "RemoveContainer" containerID="baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.051033 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 17 08:40:24 crc kubenswrapper[4966]: E1217 08:40:24.054033 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c\": container with ID starting with baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c not found: ID does not exist" containerID="baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.054078 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c"} err="failed to get container status \"baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c\": rpc error: code = NotFound desc = could not find container \"baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c\": container with ID starting with baad33d886478cd193f1947457f71c46416c7881c3536847ecd0344781b96b1c not found: ID does not exist" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.170702 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.170776 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.170805 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-scripts\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.170855 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jlfg\" (UniqueName: \"kubernetes.io/projected/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-kube-api-access-7jlfg\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.170945 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-config-data\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.170991 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.272168 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.272424 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-scripts\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.273089 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jlfg\" (UniqueName: \"kubernetes.io/projected/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-kube-api-access-7jlfg\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.273519 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-config-data\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.273640 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.273811 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.273963 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.277053 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-scripts\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.277197 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.288798 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.289634 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jlfg\" (UniqueName: \"kubernetes.io/projected/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-kube-api-access-7jlfg\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.290739 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/224f6aa1-428f-4a5f-b886-e866bbd0bb8b-config-data\") pod \"cinder-scheduler-0\" (UID: \"224f6aa1-428f-4a5f-b886-e866bbd0bb8b\") " pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.403760 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.651408 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.798474 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-combined-ca-bundle\") pod \"df474304-225d-4dcc-8b31-05646232f0ae\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.798551 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data-custom\") pod \"df474304-225d-4dcc-8b31-05646232f0ae\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.798600 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data\") pod \"df474304-225d-4dcc-8b31-05646232f0ae\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.798684 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df474304-225d-4dcc-8b31-05646232f0ae-logs\") pod \"df474304-225d-4dcc-8b31-05646232f0ae\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.798709 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k872\" (UniqueName: \"kubernetes.io/projected/df474304-225d-4dcc-8b31-05646232f0ae-kube-api-access-6k872\") pod \"df474304-225d-4dcc-8b31-05646232f0ae\" (UID: \"df474304-225d-4dcc-8b31-05646232f0ae\") " Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.803327 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df474304-225d-4dcc-8b31-05646232f0ae-logs" (OuterVolumeSpecName: "logs") pod "df474304-225d-4dcc-8b31-05646232f0ae" (UID: "df474304-225d-4dcc-8b31-05646232f0ae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.805359 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df474304-225d-4dcc-8b31-05646232f0ae-kube-api-access-6k872" (OuterVolumeSpecName: "kube-api-access-6k872") pod "df474304-225d-4dcc-8b31-05646232f0ae" (UID: "df474304-225d-4dcc-8b31-05646232f0ae"). InnerVolumeSpecName "kube-api-access-6k872". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.811338 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "df474304-225d-4dcc-8b31-05646232f0ae" (UID: "df474304-225d-4dcc-8b31-05646232f0ae"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.834116 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df474304-225d-4dcc-8b31-05646232f0ae" (UID: "df474304-225d-4dcc-8b31-05646232f0ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.846798 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec394811-ff3a-49d5-86f7-6b736748f0c3" path="/var/lib/kubelet/pods/ec394811-ff3a-49d5-86f7-6b736748f0c3/volumes" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.884443 4966 generic.go:334] "Generic (PLEG): container finished" podID="df474304-225d-4dcc-8b31-05646232f0ae" containerID="b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a" exitCode=143 Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.884666 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5489d786c6-qz2lq" event={"ID":"df474304-225d-4dcc-8b31-05646232f0ae","Type":"ContainerDied","Data":"b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a"} Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.884757 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5489d786c6-qz2lq" event={"ID":"df474304-225d-4dcc-8b31-05646232f0ae","Type":"ContainerDied","Data":"d48f5b81b22c3a5bbf1777ae8f52ecfb84e08dad6019d0925541ae512f59555e"} Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.884864 4966 scope.go:117] "RemoveContainer" containerID="b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.885261 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5489d786c6-qz2lq" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.894987 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data" (OuterVolumeSpecName: "config-data") pod "df474304-225d-4dcc-8b31-05646232f0ae" (UID: "df474304-225d-4dcc-8b31-05646232f0ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.901095 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.901217 4966 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.901320 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df474304-225d-4dcc-8b31-05646232f0ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.901381 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df474304-225d-4dcc-8b31-05646232f0ae-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.901550 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k872\" (UniqueName: \"kubernetes.io/projected/df474304-225d-4dcc-8b31-05646232f0ae-kube-api-access-6k872\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:24 crc kubenswrapper[4966]: I1217 08:40:24.978978 4966 scope.go:117] "RemoveContainer" containerID="e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864" Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.004833 4966 scope.go:117] "RemoveContainer" containerID="441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5" Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.014758 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 17 08:40:25 crc kubenswrapper[4966]: W1217 08:40:25.018944 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod224f6aa1_428f_4a5f_b886_e866bbd0bb8b.slice/crio-e6103bb5f7dd5606e1c982c6ac12b4389740bac385500dc2a3dd104747a6c2bd WatchSource:0}: Error finding container e6103bb5f7dd5606e1c982c6ac12b4389740bac385500dc2a3dd104747a6c2bd: Status 404 returned error can't find the container with id e6103bb5f7dd5606e1c982c6ac12b4389740bac385500dc2a3dd104747a6c2bd Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.150294 4966 scope.go:117] "RemoveContainer" containerID="b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a" Dec 17 08:40:25 crc kubenswrapper[4966]: E1217 08:40:25.150698 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a\": container with ID starting with b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a not found: ID does not exist" containerID="b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a" Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.150732 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a"} err="failed to get container status \"b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a\": rpc error: code = NotFound desc = could not find container \"b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a\": container with ID starting with b1f870a49615a182285a2e3cdbc339b6a9c8510630091850e94d7de619259e0a not found: ID does not exist" Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.150757 4966 scope.go:117] "RemoveContainer" containerID="e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864" Dec 17 08:40:25 crc kubenswrapper[4966]: E1217 08:40:25.151134 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864\": container with ID starting with e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864 not found: ID does not exist" containerID="e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864" Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.151161 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864"} err="failed to get container status \"e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864\": rpc error: code = NotFound desc = could not find container \"e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864\": container with ID starting with e1a2a4952d5049153918fe17a022ad4c42ea441be7cf1a26a56168550ea20864 not found: ID does not exist" Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.151175 4966 scope.go:117] "RemoveContainer" containerID="441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5" Dec 17 08:40:25 crc kubenswrapper[4966]: E1217 08:40:25.151405 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5\": container with ID starting with 441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5 not found: ID does not exist" containerID="441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5" Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.151430 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5"} err="failed to get container status \"441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5\": rpc error: code = NotFound desc = could not find container \"441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5\": container with ID starting with 441069098c993daca08dae33f68f078e4f6e1ae000696a8e75f785360121b6b5 not found: ID does not exist" Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.255756 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5489d786c6-qz2lq"] Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.270098 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5489d786c6-qz2lq"] Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.931457 4966 generic.go:334] "Generic (PLEG): container finished" podID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerID="ff860a43024f6954192167e387c152dc2b8c1a2ad2cbeaee0c6229e8eb0dd637" exitCode=0 Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.931529 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fbfdd886-69frk" event={"ID":"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017","Type":"ContainerDied","Data":"ff860a43024f6954192167e387c152dc2b8c1a2ad2cbeaee0c6229e8eb0dd637"} Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.935070 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"224f6aa1-428f-4a5f-b886-e866bbd0bb8b","Type":"ContainerStarted","Data":"3e04b0cf2db3fe9512c5dd184a68f744f46da993c27b9095c23d30449d234306"} Dec 17 08:40:25 crc kubenswrapper[4966]: I1217 08:40:25.935096 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"224f6aa1-428f-4a5f-b886-e866bbd0bb8b","Type":"ContainerStarted","Data":"e6103bb5f7dd5606e1c982c6ac12b4389740bac385500dc2a3dd104747a6c2bd"} Dec 17 08:40:26 crc kubenswrapper[4966]: I1217 08:40:26.843357 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df474304-225d-4dcc-8b31-05646232f0ae" path="/var/lib/kubelet/pods/df474304-225d-4dcc-8b31-05646232f0ae/volumes" Dec 17 08:40:26 crc kubenswrapper[4966]: I1217 08:40:26.944464 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"224f6aa1-428f-4a5f-b886-e866bbd0bb8b","Type":"ContainerStarted","Data":"b71e5781b5d5e3be920d43f4e35d9607a707294261c24f31083a1964a5601b6b"} Dec 17 08:40:27 crc kubenswrapper[4966]: I1217 08:40:27.347429 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-fbfdd886-69frk" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.156:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.156:8443: connect: connection refused" Dec 17 08:40:27 crc kubenswrapper[4966]: I1217 08:40:27.825052 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 17 08:40:28 crc kubenswrapper[4966]: I1217 08:40:28.154500 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:40:28 crc kubenswrapper[4966]: I1217 08:40:28.155710 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-595bcd9766-wkh6b" Dec 17 08:40:28 crc kubenswrapper[4966]: I1217 08:40:28.173719 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.173700912 podStartE2EDuration="5.173700912s" podCreationTimestamp="2025-12-17 08:40:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:26.980115806 +0000 UTC m=+1162.525185748" watchObservedRunningTime="2025-12-17 08:40:28.173700912 +0000 UTC m=+1163.718770854" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.104110 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-879977d5b-rwf8b" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.404988 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.535274 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 17 08:40:29 crc kubenswrapper[4966]: E1217 08:40:29.535646 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.535657 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api" Dec 17 08:40:29 crc kubenswrapper[4966]: E1217 08:40:29.535666 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.535672 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" Dec 17 08:40:29 crc kubenswrapper[4966]: E1217 08:40:29.535681 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.535687 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" Dec 17 08:40:29 crc kubenswrapper[4966]: E1217 08:40:29.535716 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.535722 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.535907 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.535917 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.535931 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.535944 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="df474304-225d-4dcc-8b31-05646232f0ae" containerName="barbican-api-log" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.536466 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.542115 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.542153 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.542323 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-jks9n" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.556523 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.596075 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/220e53bd-2342-4bed-8e26-487e0859350b-openstack-config-secret\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.596115 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjc6b\" (UniqueName: \"kubernetes.io/projected/220e53bd-2342-4bed-8e26-487e0859350b-kube-api-access-jjc6b\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.596218 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/220e53bd-2342-4bed-8e26-487e0859350b-openstack-config\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.596239 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220e53bd-2342-4bed-8e26-487e0859350b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.698347 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/220e53bd-2342-4bed-8e26-487e0859350b-openstack-config\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.698685 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220e53bd-2342-4bed-8e26-487e0859350b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.698820 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/220e53bd-2342-4bed-8e26-487e0859350b-openstack-config-secret\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.698844 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjc6b\" (UniqueName: \"kubernetes.io/projected/220e53bd-2342-4bed-8e26-487e0859350b-kube-api-access-jjc6b\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.699425 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/220e53bd-2342-4bed-8e26-487e0859350b-openstack-config\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.705611 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220e53bd-2342-4bed-8e26-487e0859350b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.706047 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/220e53bd-2342-4bed-8e26-487e0859350b-openstack-config-secret\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.719748 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjc6b\" (UniqueName: \"kubernetes.io/projected/220e53bd-2342-4bed-8e26-487e0859350b-kube-api-access-jjc6b\") pod \"openstackclient\" (UID: \"220e53bd-2342-4bed-8e26-487e0859350b\") " pod="openstack/openstackclient" Dec 17 08:40:29 crc kubenswrapper[4966]: I1217 08:40:29.859946 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 17 08:40:30 crc kubenswrapper[4966]: I1217 08:40:30.457077 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 17 08:40:30 crc kubenswrapper[4966]: I1217 08:40:30.977962 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"220e53bd-2342-4bed-8e26-487e0859350b","Type":"ContainerStarted","Data":"bee54604f131bbb5b50154e1c14f93dc3b578ec1dbaf47745bb71230943b0b13"} Dec 17 08:40:31 crc kubenswrapper[4966]: I1217 08:40:31.370528 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 17 08:40:34 crc kubenswrapper[4966]: I1217 08:40:34.734370 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.430939 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-b67cfd874-bklxp"] Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.433040 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.461735 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.464823 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.465266 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-l4hjd" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.471684 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.479223 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7c6p\" (UniqueName: \"kubernetes.io/projected/f33826e0-0269-4b99-9dce-fcaf62f67bed-kube-api-access-k7c6p\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.481599 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data-custom\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.482095 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-combined-ca-bundle\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.514936 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-b67cfd874-bklxp"] Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.586928 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.587010 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7c6p\" (UniqueName: \"kubernetes.io/projected/f33826e0-0269-4b99-9dce-fcaf62f67bed-kube-api-access-k7c6p\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.587042 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data-custom\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.587069 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-combined-ca-bundle\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.600638 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-combined-ca-bundle\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.608539 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data-custom\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.610785 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.635206 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7c6p\" (UniqueName: \"kubernetes.io/projected/f33826e0-0269-4b99-9dce-fcaf62f67bed-kube-api-access-k7c6p\") pod \"heat-engine-b67cfd874-bklxp\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.762688 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.803560 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c4f4f4d89-wmvz6"] Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.805186 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.826382 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7ccdc8f9cd-5gvvt"] Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.827557 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.838236 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.867680 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c4f4f4d89-wmvz6"] Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.878893 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7ccdc8f9cd-5gvvt"] Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.967669 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6c76d785bf-vjmcl"] Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.968800 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.985666 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.992511 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-swift-storage-0\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.992580 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqzcb\" (UniqueName: \"kubernetes.io/projected/7fb54742-7249-4b48-a918-1bfab858a763-kube-api-access-xqzcb\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.992642 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q258b\" (UniqueName: \"kubernetes.io/projected/6c618545-5898-4e00-9ed3-5f34f2ad77c8-kube-api-access-q258b\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.992670 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-sb\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.992730 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-config\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.992784 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-svc\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:36 crc kubenswrapper[4966]: I1217 08:40:36.992823 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-nb\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.004019 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data-custom\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.004193 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-combined-ca-bundle\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.004241 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.034046 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6c76d785bf-vjmcl"] Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109234 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqzcb\" (UniqueName: \"kubernetes.io/projected/7fb54742-7249-4b48-a918-1bfab858a763-kube-api-access-xqzcb\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109283 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdmsv\" (UniqueName: \"kubernetes.io/projected/0b528e6e-b726-4895-910c-8544dcf26417-kube-api-access-vdmsv\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109340 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q258b\" (UniqueName: \"kubernetes.io/projected/6c618545-5898-4e00-9ed3-5f34f2ad77c8-kube-api-access-q258b\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109369 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-sb\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109413 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-config\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109446 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-svc\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109480 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-nb\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109515 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data-custom\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109543 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109583 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data-custom\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109618 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-combined-ca-bundle\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109650 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-combined-ca-bundle\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109678 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.109727 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-swift-storage-0\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.119819 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data-custom\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.121720 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-sb\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.122686 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-svc\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.132624 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-config\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.133282 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-swift-storage-0\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.133318 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-nb\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.141304 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.158704 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-combined-ca-bundle\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.172028 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqzcb\" (UniqueName: \"kubernetes.io/projected/7fb54742-7249-4b48-a918-1bfab858a763-kube-api-access-xqzcb\") pod \"dnsmasq-dns-c4f4f4d89-wmvz6\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.172110 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q258b\" (UniqueName: \"kubernetes.io/projected/6c618545-5898-4e00-9ed3-5f34f2ad77c8-kube-api-access-q258b\") pod \"heat-cfnapi-7ccdc8f9cd-5gvvt\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.210778 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdmsv\" (UniqueName: \"kubernetes.io/projected/0b528e6e-b726-4895-910c-8544dcf26417-kube-api-access-vdmsv\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.210896 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data-custom\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.210920 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.210956 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-combined-ca-bundle\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.216475 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-combined-ca-bundle\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.217535 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.220814 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data-custom\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.270069 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdmsv\" (UniqueName: \"kubernetes.io/projected/0b528e6e-b726-4895-910c-8544dcf26417-kube-api-access-vdmsv\") pod \"heat-api-6c76d785bf-vjmcl\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.303391 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.346560 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-fbfdd886-69frk" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.156:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.156:8443: connect: connection refused" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.450693 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:37 crc kubenswrapper[4966]: I1217 08:40:37.466980 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:40 crc kubenswrapper[4966]: I1217 08:40:40.867144 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-xlrkm"] Dec 17 08:40:40 crc kubenswrapper[4966]: I1217 08:40:40.868485 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:40 crc kubenswrapper[4966]: I1217 08:40:40.888252 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xlrkm"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.003913 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55hq4\" (UniqueName: \"kubernetes.io/projected/0644984a-4f00-4b7a-9a05-d9e43097a7c4-kube-api-access-55hq4\") pod \"nova-api-db-create-xlrkm\" (UID: \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\") " pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.003971 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0644984a-4f00-4b7a-9a05-d9e43097a7c4-operator-scripts\") pod \"nova-api-db-create-xlrkm\" (UID: \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\") " pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.076227 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-7d41-account-create-update-x4fn4"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.077589 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:41 crc kubenswrapper[4966]: W1217 08:40:41.087417 4966 reflector.go:561] object-"openstack"/"nova-api-db-secret": failed to list *v1.Secret: secrets "nova-api-db-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 17 08:40:41 crc kubenswrapper[4966]: E1217 08:40:41.087455 4966 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"nova-api-db-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"nova-api-db-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.094634 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-spxpd"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.095885 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.104357 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7d41-account-create-update-x4fn4"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.105484 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55hq4\" (UniqueName: \"kubernetes.io/projected/0644984a-4f00-4b7a-9a05-d9e43097a7c4-kube-api-access-55hq4\") pod \"nova-api-db-create-xlrkm\" (UID: \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\") " pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.105523 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0644984a-4f00-4b7a-9a05-d9e43097a7c4-operator-scripts\") pod \"nova-api-db-create-xlrkm\" (UID: \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\") " pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.107357 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0644984a-4f00-4b7a-9a05-d9e43097a7c4-operator-scripts\") pod \"nova-api-db-create-xlrkm\" (UID: \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\") " pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.116549 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-spxpd"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.177153 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55hq4\" (UniqueName: \"kubernetes.io/projected/0644984a-4f00-4b7a-9a05-d9e43097a7c4-kube-api-access-55hq4\") pod \"nova-api-db-create-xlrkm\" (UID: \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\") " pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.211747 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b302878-13e7-4967-aaf1-58292a7237e5-operator-scripts\") pod \"nova-cell0-db-create-spxpd\" (UID: \"2b302878-13e7-4967-aaf1-58292a7237e5\") " pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.211813 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr7sv\" (UniqueName: \"kubernetes.io/projected/2b302878-13e7-4967-aaf1-58292a7237e5-kube-api-access-wr7sv\") pod \"nova-cell0-db-create-spxpd\" (UID: \"2b302878-13e7-4967-aaf1-58292a7237e5\") " pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.211829 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s9vw\" (UniqueName: \"kubernetes.io/projected/ce426531-9f5f-4c9d-8bea-e22d666fe95d-kube-api-access-7s9vw\") pod \"nova-api-7d41-account-create-update-x4fn4\" (UID: \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\") " pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.211930 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce426531-9f5f-4c9d-8bea-e22d666fe95d-operator-scripts\") pod \"nova-api-7d41-account-create-update-x4fn4\" (UID: \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\") " pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.254975 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.283236 4966 generic.go:334] "Generic (PLEG): container finished" podID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerID="5493e36eaf0a74e0ee898b72a7b70f252fb9beca3c6da8ed9c11248832a6fc69" exitCode=137 Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.283289 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerDied","Data":"5493e36eaf0a74e0ee898b72a7b70f252fb9beca3c6da8ed9c11248832a6fc69"} Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.304908 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-c4llv"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.306202 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.314368 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b302878-13e7-4967-aaf1-58292a7237e5-operator-scripts\") pod \"nova-cell0-db-create-spxpd\" (UID: \"2b302878-13e7-4967-aaf1-58292a7237e5\") " pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.314414 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr7sv\" (UniqueName: \"kubernetes.io/projected/2b302878-13e7-4967-aaf1-58292a7237e5-kube-api-access-wr7sv\") pod \"nova-cell0-db-create-spxpd\" (UID: \"2b302878-13e7-4967-aaf1-58292a7237e5\") " pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.314431 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s9vw\" (UniqueName: \"kubernetes.io/projected/ce426531-9f5f-4c9d-8bea-e22d666fe95d-kube-api-access-7s9vw\") pod \"nova-api-7d41-account-create-update-x4fn4\" (UID: \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\") " pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.314513 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce426531-9f5f-4c9d-8bea-e22d666fe95d-operator-scripts\") pod \"nova-api-7d41-account-create-update-x4fn4\" (UID: \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\") " pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.315615 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce426531-9f5f-4c9d-8bea-e22d666fe95d-operator-scripts\") pod \"nova-api-7d41-account-create-update-x4fn4\" (UID: \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\") " pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.316745 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b302878-13e7-4967-aaf1-58292a7237e5-operator-scripts\") pod \"nova-cell0-db-create-spxpd\" (UID: \"2b302878-13e7-4967-aaf1-58292a7237e5\") " pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.371937 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-c4llv"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.379047 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr7sv\" (UniqueName: \"kubernetes.io/projected/2b302878-13e7-4967-aaf1-58292a7237e5-kube-api-access-wr7sv\") pod \"nova-cell0-db-create-spxpd\" (UID: \"2b302878-13e7-4967-aaf1-58292a7237e5\") " pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.389620 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s9vw\" (UniqueName: \"kubernetes.io/projected/ce426531-9f5f-4c9d-8bea-e22d666fe95d-kube-api-access-7s9vw\") pod \"nova-api-7d41-account-create-update-x4fn4\" (UID: \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\") " pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.432098 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b05f9d6-237c-4cb0-91b9-505715528c7d-operator-scripts\") pod \"nova-cell1-db-create-c4llv\" (UID: \"6b05f9d6-237c-4cb0-91b9-505715528c7d\") " pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.432256 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ckl4\" (UniqueName: \"kubernetes.io/projected/6b05f9d6-237c-4cb0-91b9-505715528c7d-kube-api-access-2ckl4\") pod \"nova-cell1-db-create-c4llv\" (UID: \"6b05f9d6-237c-4cb0-91b9-505715528c7d\") " pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.440524 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.469615 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.475949 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-67a5-account-create-update-qdhbr"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.477135 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.480262 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.535785 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b05f9d6-237c-4cb0-91b9-505715528c7d-operator-scripts\") pod \"nova-cell1-db-create-c4llv\" (UID: \"6b05f9d6-237c-4cb0-91b9-505715528c7d\") " pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.535843 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdhrw\" (UniqueName: \"kubernetes.io/projected/fed2ac76-ebb0-481f-82e1-55397f658c99-kube-api-access-tdhrw\") pod \"nova-cell0-67a5-account-create-update-qdhbr\" (UID: \"fed2ac76-ebb0-481f-82e1-55397f658c99\") " pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.536602 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed2ac76-ebb0-481f-82e1-55397f658c99-operator-scripts\") pod \"nova-cell0-67a5-account-create-update-qdhbr\" (UID: \"fed2ac76-ebb0-481f-82e1-55397f658c99\") " pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.536637 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ckl4\" (UniqueName: \"kubernetes.io/projected/6b05f9d6-237c-4cb0-91b9-505715528c7d-kube-api-access-2ckl4\") pod \"nova-cell1-db-create-c4llv\" (UID: \"6b05f9d6-237c-4cb0-91b9-505715528c7d\") " pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.537502 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b05f9d6-237c-4cb0-91b9-505715528c7d-operator-scripts\") pod \"nova-cell1-db-create-c4llv\" (UID: \"6b05f9d6-237c-4cb0-91b9-505715528c7d\") " pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.573116 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-67a5-account-create-update-qdhbr"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.600569 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ckl4\" (UniqueName: \"kubernetes.io/projected/6b05f9d6-237c-4cb0-91b9-505715528c7d-kube-api-access-2ckl4\") pod \"nova-cell1-db-create-c4llv\" (UID: \"6b05f9d6-237c-4cb0-91b9-505715528c7d\") " pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.640314 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed2ac76-ebb0-481f-82e1-55397f658c99-operator-scripts\") pod \"nova-cell0-67a5-account-create-update-qdhbr\" (UID: \"fed2ac76-ebb0-481f-82e1-55397f658c99\") " pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.640551 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdhrw\" (UniqueName: \"kubernetes.io/projected/fed2ac76-ebb0-481f-82e1-55397f658c99-kube-api-access-tdhrw\") pod \"nova-cell0-67a5-account-create-update-qdhbr\" (UID: \"fed2ac76-ebb0-481f-82e1-55397f658c99\") " pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.648680 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed2ac76-ebb0-481f-82e1-55397f658c99-operator-scripts\") pod \"nova-cell0-67a5-account-create-update-qdhbr\" (UID: \"fed2ac76-ebb0-481f-82e1-55397f658c99\") " pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.660059 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-e4a6-account-create-update-q8vbr"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.672996 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.698575 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.739415 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdhrw\" (UniqueName: \"kubernetes.io/projected/fed2ac76-ebb0-481f-82e1-55397f658c99-kube-api-access-tdhrw\") pod \"nova-cell0-67a5-account-create-update-qdhbr\" (UID: \"fed2ac76-ebb0-481f-82e1-55397f658c99\") " pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.760326 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e4a6-account-create-update-q8vbr"] Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.761502 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.832256 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.849864 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9ea912d-6078-4ab1-8188-f1e2927eb12a-operator-scripts\") pod \"nova-cell1-e4a6-account-create-update-q8vbr\" (UID: \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\") " pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.850712 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rn2n\" (UniqueName: \"kubernetes.io/projected/c9ea912d-6078-4ab1-8188-f1e2927eb12a-kube-api-access-2rn2n\") pod \"nova-cell1-e4a6-account-create-update-q8vbr\" (UID: \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\") " pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.952705 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rn2n\" (UniqueName: \"kubernetes.io/projected/c9ea912d-6078-4ab1-8188-f1e2927eb12a-kube-api-access-2rn2n\") pod \"nova-cell1-e4a6-account-create-update-q8vbr\" (UID: \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\") " pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.952851 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9ea912d-6078-4ab1-8188-f1e2927eb12a-operator-scripts\") pod \"nova-cell1-e4a6-account-create-update-q8vbr\" (UID: \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\") " pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.953546 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9ea912d-6078-4ab1-8188-f1e2927eb12a-operator-scripts\") pod \"nova-cell1-e4a6-account-create-update-q8vbr\" (UID: \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\") " pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:41 crc kubenswrapper[4966]: I1217 08:40:41.973935 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rn2n\" (UniqueName: \"kubernetes.io/projected/c9ea912d-6078-4ab1-8188-f1e2927eb12a-kube-api-access-2rn2n\") pod \"nova-cell1-e4a6-account-create-update-q8vbr\" (UID: \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\") " pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.028079 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5679cd5d4c-bstjs"] Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.033264 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.045039 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.045298 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.045419 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.095427 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.107155 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5679cd5d4c-bstjs"] Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.174632 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-config-data\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.188319 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-combined-ca-bundle\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.188379 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0333b211-1df2-45b3-9f53-bfcc921f98cc-etc-swift\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.188480 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-internal-tls-certs\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.188514 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0333b211-1df2-45b3-9f53-bfcc921f98cc-run-httpd\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.188580 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fhs2\" (UniqueName: \"kubernetes.io/projected/0333b211-1df2-45b3-9f53-bfcc921f98cc-kube-api-access-8fhs2\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.188631 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-public-tls-certs\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.188686 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0333b211-1df2-45b3-9f53-bfcc921f98cc-log-httpd\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.291152 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-config-data\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.291200 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-combined-ca-bundle\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.291221 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0333b211-1df2-45b3-9f53-bfcc921f98cc-etc-swift\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.291258 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-internal-tls-certs\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.291280 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0333b211-1df2-45b3-9f53-bfcc921f98cc-run-httpd\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.291316 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fhs2\" (UniqueName: \"kubernetes.io/projected/0333b211-1df2-45b3-9f53-bfcc921f98cc-kube-api-access-8fhs2\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.291344 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-public-tls-certs\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.291373 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0333b211-1df2-45b3-9f53-bfcc921f98cc-log-httpd\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.291916 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0333b211-1df2-45b3-9f53-bfcc921f98cc-log-httpd\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.292395 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0333b211-1df2-45b3-9f53-bfcc921f98cc-run-httpd\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.296791 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-config-data\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.299489 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-public-tls-certs\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.304008 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-internal-tls-certs\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.308723 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0333b211-1df2-45b3-9f53-bfcc921f98cc-combined-ca-bundle\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.313063 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0333b211-1df2-45b3-9f53-bfcc921f98cc-etc-swift\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.316429 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.324972 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fhs2\" (UniqueName: \"kubernetes.io/projected/0333b211-1df2-45b3-9f53-bfcc921f98cc-kube-api-access-8fhs2\") pod \"swift-proxy-5679cd5d4c-bstjs\" (UID: \"0333b211-1df2-45b3-9f53-bfcc921f98cc\") " pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:42 crc kubenswrapper[4966]: I1217 08:40:42.446724 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.016911 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7f66c49f7b-4992d"] Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.018214 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.058182 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-68976d496b-zns8c"] Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.059467 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.074384 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7f66c49f7b-4992d"] Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.130805 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-combined-ca-bundle\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.131415 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg2rn\" (UniqueName: \"kubernetes.io/projected/f1e350c4-e807-4015-8398-42bad206482f-kube-api-access-fg2rn\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.131496 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data-custom\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.131624 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/707f4037-3896-482d-85a8-3b0fc081045d-config-data-custom\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.131682 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707f4037-3896-482d-85a8-3b0fc081045d-combined-ca-bundle\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.131743 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/707f4037-3896-482d-85a8-3b0fc081045d-config-data\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.131931 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frsbn\" (UniqueName: \"kubernetes.io/projected/707f4037-3896-482d-85a8-3b0fc081045d-kube-api-access-frsbn\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.131970 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.164932 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-68976d496b-zns8c"] Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.200923 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7668494dc4-2sqhx"] Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.202112 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.213666 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7668494dc4-2sqhx"] Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.233259 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frsbn\" (UniqueName: \"kubernetes.io/projected/707f4037-3896-482d-85a8-3b0fc081045d-kube-api-access-frsbn\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.233303 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.233373 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-combined-ca-bundle\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.233424 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg2rn\" (UniqueName: \"kubernetes.io/projected/f1e350c4-e807-4015-8398-42bad206482f-kube-api-access-fg2rn\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.233452 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data-custom\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.233477 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/707f4037-3896-482d-85a8-3b0fc081045d-config-data-custom\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.233494 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707f4037-3896-482d-85a8-3b0fc081045d-combined-ca-bundle\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.233519 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/707f4037-3896-482d-85a8-3b0fc081045d-config-data\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.240309 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data-custom\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.241156 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/707f4037-3896-482d-85a8-3b0fc081045d-config-data\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.244636 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707f4037-3896-482d-85a8-3b0fc081045d-combined-ca-bundle\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.245845 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.252496 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg2rn\" (UniqueName: \"kubernetes.io/projected/f1e350c4-e807-4015-8398-42bad206482f-kube-api-access-fg2rn\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.260692 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frsbn\" (UniqueName: \"kubernetes.io/projected/707f4037-3896-482d-85a8-3b0fc081045d-kube-api-access-frsbn\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.264445 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-combined-ca-bundle\") pod \"heat-api-68976d496b-zns8c\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.270495 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/707f4037-3896-482d-85a8-3b0fc081045d-config-data-custom\") pod \"heat-engine-7f66c49f7b-4992d\" (UID: \"707f4037-3896-482d-85a8-3b0fc081045d\") " pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.335530 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmpsh\" (UniqueName: \"kubernetes.io/projected/cdfc5775-0ec0-433c-9774-0e6a56273b8e-kube-api-access-vmpsh\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.335622 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.335724 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data-custom\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.335809 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-combined-ca-bundle\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.348423 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.384270 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.437569 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.437685 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data-custom\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.437739 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-combined-ca-bundle\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.437799 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmpsh\" (UniqueName: \"kubernetes.io/projected/cdfc5775-0ec0-433c-9774-0e6a56273b8e-kube-api-access-vmpsh\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.444656 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data-custom\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.445070 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.454452 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-combined-ca-bundle\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.455825 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmpsh\" (UniqueName: \"kubernetes.io/projected/cdfc5775-0ec0-433c-9774-0e6a56273b8e-kube-api-access-vmpsh\") pod \"heat-cfnapi-7668494dc4-2sqhx\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:44 crc kubenswrapper[4966]: I1217 08:40:44.526377 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:46 crc kubenswrapper[4966]: E1217 08:40:46.014790 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-openstackclient:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:40:46 crc kubenswrapper[4966]: E1217 08:40:46.015158 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-openstackclient:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:40:46 crc kubenswrapper[4966]: E1217 08:40:46.015287 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-openstackclient:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbfh98h648h666hbbh8bh56bh545h557h57dh5dch5b6h676h7bhd4hc5h594h5bh66bhf8h695h79hf4hbbh67ch5bhb5h58h84h58fh548h55q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jjc6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(220e53bd-2342-4bed-8e26-487e0859350b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:40:46 crc kubenswrapper[4966]: E1217 08:40:46.020551 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="220e53bd-2342-4bed-8e26-487e0859350b" Dec 17 08:40:46 crc kubenswrapper[4966]: E1217 08:40:46.464243 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.156:5001/podified-antelope-centos9/openstack-openstackclient:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/openstackclient" podUID="220e53bd-2342-4bed-8e26-487e0859350b" Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.725306 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.895739 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xlrkm"] Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.931702 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-config-data\") pod \"b26b94bc-5f74-4321-8e33-28b28b94ac50\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.940316 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf4xx\" (UniqueName: \"kubernetes.io/projected/b26b94bc-5f74-4321-8e33-28b28b94ac50-kube-api-access-cf4xx\") pod \"b26b94bc-5f74-4321-8e33-28b28b94ac50\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.940357 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-log-httpd\") pod \"b26b94bc-5f74-4321-8e33-28b28b94ac50\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.940411 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-run-httpd\") pod \"b26b94bc-5f74-4321-8e33-28b28b94ac50\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.940440 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-scripts\") pod \"b26b94bc-5f74-4321-8e33-28b28b94ac50\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.940473 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-sg-core-conf-yaml\") pod \"b26b94bc-5f74-4321-8e33-28b28b94ac50\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.940543 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-combined-ca-bundle\") pod \"b26b94bc-5f74-4321-8e33-28b28b94ac50\" (UID: \"b26b94bc-5f74-4321-8e33-28b28b94ac50\") " Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.950535 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b26b94bc-5f74-4321-8e33-28b28b94ac50" (UID: "b26b94bc-5f74-4321-8e33-28b28b94ac50"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.951113 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b26b94bc-5f74-4321-8e33-28b28b94ac50" (UID: "b26b94bc-5f74-4321-8e33-28b28b94ac50"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.955176 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-scripts" (OuterVolumeSpecName: "scripts") pod "b26b94bc-5f74-4321-8e33-28b28b94ac50" (UID: "b26b94bc-5f74-4321-8e33-28b28b94ac50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:46 crc kubenswrapper[4966]: I1217 08:40:46.966146 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b26b94bc-5f74-4321-8e33-28b28b94ac50-kube-api-access-cf4xx" (OuterVolumeSpecName: "kube-api-access-cf4xx") pod "b26b94bc-5f74-4321-8e33-28b28b94ac50" (UID: "b26b94bc-5f74-4321-8e33-28b28b94ac50"). InnerVolumeSpecName "kube-api-access-cf4xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.061360 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf4xx\" (UniqueName: \"kubernetes.io/projected/b26b94bc-5f74-4321-8e33-28b28b94ac50-kube-api-access-cf4xx\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.061394 4966 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.061404 4966 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b26b94bc-5f74-4321-8e33-28b28b94ac50-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.061412 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.079850 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b26b94bc-5f74-4321-8e33-28b28b94ac50" (UID: "b26b94bc-5f74-4321-8e33-28b28b94ac50"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.168815 4966 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.187448 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b26b94bc-5f74-4321-8e33-28b28b94ac50" (UID: "b26b94bc-5f74-4321-8e33-28b28b94ac50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.248282 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6c76d785bf-vjmcl"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.277240 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.279345 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7ccdc8f9cd-5gvvt"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.309762 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6c6f84469f-k22bc"] Dec 17 08:40:47 crc kubenswrapper[4966]: E1217 08:40:47.310411 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="proxy-httpd" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.310423 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="proxy-httpd" Dec 17 08:40:47 crc kubenswrapper[4966]: E1217 08:40:47.310436 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="ceilometer-notification-agent" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.310442 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="ceilometer-notification-agent" Dec 17 08:40:47 crc kubenswrapper[4966]: E1217 08:40:47.310453 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="sg-core" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.310459 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="sg-core" Dec 17 08:40:47 crc kubenswrapper[4966]: E1217 08:40:47.310473 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="ceilometer-central-agent" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.310479 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="ceilometer-central-agent" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.310654 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="proxy-httpd" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.310667 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="ceilometer-notification-agent" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.310676 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="sg-core" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.310689 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" containerName="ceilometer-central-agent" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.312121 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.337440 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.337827 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.339337 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6cb857dc54-5f9tj"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.340528 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.370634 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.371166 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.376442 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-fbfdd886-69frk" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.156:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.156:8443: connect: connection refused" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.376529 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.381889 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-config-data-custom\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.381953 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2hk8\" (UniqueName: \"kubernetes.io/projected/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-kube-api-access-p2hk8\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382007 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-config-data\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382044 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsnpq\" (UniqueName: \"kubernetes.io/projected/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-kube-api-access-lsnpq\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382075 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-public-tls-certs\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382151 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-config-data-custom\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382184 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-config-data\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382231 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-combined-ca-bundle\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382262 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-internal-tls-certs\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382327 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-combined-ca-bundle\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382365 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-public-tls-certs\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.382437 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-internal-tls-certs\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.433246 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6c6f84469f-k22bc"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498272 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-config-data\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498359 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsnpq\" (UniqueName: \"kubernetes.io/projected/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-kube-api-access-lsnpq\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498386 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-public-tls-certs\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498443 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-config-data-custom\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498466 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-config-data\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498499 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-combined-ca-bundle\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498525 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-internal-tls-certs\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498570 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-combined-ca-bundle\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498598 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-public-tls-certs\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.498646 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-internal-tls-certs\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.511866 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-config-data-custom\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.511936 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2hk8\" (UniqueName: \"kubernetes.io/projected/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-kube-api-access-p2hk8\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.565925 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6cb857dc54-5f9tj"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.571363 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-config-data" (OuterVolumeSpecName: "config-data") pod "b26b94bc-5f74-4321-8e33-28b28b94ac50" (UID: "b26b94bc-5f74-4321-8e33-28b28b94ac50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.573563 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2hk8\" (UniqueName: \"kubernetes.io/projected/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-kube-api-access-p2hk8\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: W1217 08:40:47.578687 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b528e6e_b726_4895_910c_8544dcf26417.slice/crio-17e8cfd2a7271403fecef231c419eec9efda78d56d809247e0310b9c0b75cbdc WatchSource:0}: Error finding container 17e8cfd2a7271403fecef231c419eec9efda78d56d809247e0310b9c0b75cbdc: Status 404 returned error can't find the container with id 17e8cfd2a7271403fecef231c419eec9efda78d56d809247e0310b9c0b75cbdc Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.583738 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-public-tls-certs\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.587304 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-config-data\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.587582 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-combined-ca-bundle\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.587810 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-internal-tls-certs\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.603642 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-public-tls-certs\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.610987 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-config-data-custom\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.615133 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-config-data-custom\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.619415 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc264f63-3c8c-47d7-851b-7fcbcac72b5e-combined-ca-bundle\") pod \"heat-cfnapi-6c6f84469f-k22bc\" (UID: \"bc264f63-3c8c-47d7-851b-7fcbcac72b5e\") " pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.620316 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b26b94bc-5f74-4321-8e33-28b28b94ac50","Type":"ContainerDied","Data":"92447e11a1bfc272ca2bb89e988e45dc5b645f685826b7f261b07e71b5b73414"} Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.620405 4966 scope.go:117] "RemoveContainer" containerID="5493e36eaf0a74e0ee898b72a7b70f252fb9beca3c6da8ed9c11248832a6fc69" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.620596 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.622075 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-internal-tls-certs\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.625897 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsnpq\" (UniqueName: \"kubernetes.io/projected/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-kube-api-access-lsnpq\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.628479 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a20b1b5-aab0-4f67-ad55-023e9c908dd6-config-data\") pod \"heat-api-6cb857dc54-5f9tj\" (UID: \"1a20b1b5-aab0-4f67-ad55-023e9c908dd6\") " pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.636800 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xlrkm" event={"ID":"0644984a-4f00-4b7a-9a05-d9e43097a7c4","Type":"ContainerStarted","Data":"4ed04c2942e721e99e697deb49ab147665bc91012a929f3b49cca66e560f18f1"} Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.644839 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-c4llv"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.673481 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b26b94bc-5f74-4321-8e33-28b28b94ac50-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.694810 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6c76d785bf-vjmcl"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.767023 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.803445 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.819595 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.819602 4966 scope.go:117] "RemoveContainer" containerID="8dbc5e0821630f3a48a51d25a09ad40619167d96ffdaa221adf907da179071e2" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.850996 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.863529 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.866177 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.868097 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.872047 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.888965 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.916971 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7668494dc4-2sqhx"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.921758 4966 scope.go:117] "RemoveContainer" containerID="7bddea5ae658bbc591e4d75963b5c75b508219d0c0ae56659eabe943b39f9767" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.930418 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-b67cfd874-bklxp"] Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.981206 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-scripts\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.981264 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.981307 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.981356 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-log-httpd\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.981392 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-config-data\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.981429 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mfcd\" (UniqueName: \"kubernetes.io/projected/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-kube-api-access-5mfcd\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.981493 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-run-httpd\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:47 crc kubenswrapper[4966]: I1217 08:40:47.987148 4966 scope.go:117] "RemoveContainer" containerID="5a3ca13748e8803e6e39f736c666c4d4e73383bc2acc5ebdc79d6f5a01c7c2fd" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.083510 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.083799 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-log-httpd\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.083835 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-config-data\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.083880 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mfcd\" (UniqueName: \"kubernetes.io/projected/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-kube-api-access-5mfcd\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.083904 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-run-httpd\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.083961 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-scripts\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.083985 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.088162 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-log-httpd\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.088520 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.091830 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-config-data\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.092169 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-run-httpd\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.113372 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mfcd\" (UniqueName: \"kubernetes.io/projected/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-kube-api-access-5mfcd\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.120863 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-scripts\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.134344 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.220696 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.570774 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-e4a6-account-create-update-q8vbr"] Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.612153 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7d41-account-create-update-x4fn4"] Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.646828 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-68976d496b-zns8c"] Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.682986 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-spxpd"] Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.736563 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-67a5-account-create-update-qdhbr"] Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.793561 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c4f4f4d89-wmvz6"] Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.824521 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-b67cfd874-bklxp" event={"ID":"f33826e0-0269-4b99-9dce-fcaf62f67bed","Type":"ContainerStarted","Data":"e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6"} Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.824580 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-b67cfd874-bklxp" event={"ID":"f33826e0-0269-4b99-9dce-fcaf62f67bed","Type":"ContainerStarted","Data":"d43ee98b23617cef894d5dacee3a49ec79944953e7bdcdbfe07c8d8609c2f57e"} Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.824632 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.827375 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" event={"ID":"c9ea912d-6078-4ab1-8188-f1e2927eb12a","Type":"ContainerStarted","Data":"2d391d3f287f5b44e7cf194a14bc5854f51435b9a721415cf886fbdcae260051"} Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.883126 4966 generic.go:334] "Generic (PLEG): container finished" podID="6b05f9d6-237c-4cb0-91b9-505715528c7d" containerID="ffad271bb68bd78bf50f98a591a054eecfb0aacaec261acef467f5c1dd4af4ca" exitCode=0 Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.896997 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-b67cfd874-bklxp" podStartSLOduration=12.896973111 podStartE2EDuration="12.896973111s" podCreationTimestamp="2025-12-17 08:40:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:48.868258353 +0000 UTC m=+1184.413328295" watchObservedRunningTime="2025-12-17 08:40:48.896973111 +0000 UTC m=+1184.442043053" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.942633 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b26b94bc-5f74-4321-8e33-28b28b94ac50" path="/var/lib/kubelet/pods/b26b94bc-5f74-4321-8e33-28b28b94ac50/volumes" Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.948796 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c4llv" event={"ID":"6b05f9d6-237c-4cb0-91b9-505715528c7d","Type":"ContainerDied","Data":"ffad271bb68bd78bf50f98a591a054eecfb0aacaec261acef467f5c1dd4af4ca"} Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.948846 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5679cd5d4c-bstjs"] Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.948863 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c4llv" event={"ID":"6b05f9d6-237c-4cb0-91b9-505715528c7d","Type":"ContainerStarted","Data":"3532ff265744c4eec436627e330a9c4913e355a8d467950afd2e251654b91608"} Dec 17 08:40:48 crc kubenswrapper[4966]: I1217 08:40:48.948888 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" event={"ID":"cdfc5775-0ec0-433c-9774-0e6a56273b8e","Type":"ContainerStarted","Data":"0c0bd0204551f09f6fdbb12ede1f58faf732ce8c0db0882c3e327efcbfdfe3d1"} Dec 17 08:40:49 crc kubenswrapper[4966]: I1217 08:40:49.048102 4966 generic.go:334] "Generic (PLEG): container finished" podID="0644984a-4f00-4b7a-9a05-d9e43097a7c4" containerID="c0754d5a0deea891d2ebd07e50c4cd12a376080dc1b6c8a5a2a5342077161320" exitCode=0 Dec 17 08:40:49 crc kubenswrapper[4966]: I1217 08:40:49.048198 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xlrkm" event={"ID":"0644984a-4f00-4b7a-9a05-d9e43097a7c4","Type":"ContainerDied","Data":"c0754d5a0deea891d2ebd07e50c4cd12a376080dc1b6c8a5a2a5342077161320"} Dec 17 08:40:49 crc kubenswrapper[4966]: I1217 08:40:49.088214 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6c76d785bf-vjmcl" event={"ID":"0b528e6e-b726-4895-910c-8544dcf26417","Type":"ContainerStarted","Data":"17e8cfd2a7271403fecef231c419eec9efda78d56d809247e0310b9c0b75cbdc"} Dec 17 08:40:49 crc kubenswrapper[4966]: I1217 08:40:49.143572 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7f66c49f7b-4992d"] Dec 17 08:40:49 crc kubenswrapper[4966]: I1217 08:40:49.364218 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7ccdc8f9cd-5gvvt"] Dec 17 08:40:49 crc kubenswrapper[4966]: I1217 08:40:49.397027 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6cb857dc54-5f9tj"] Dec 17 08:40:49 crc kubenswrapper[4966]: I1217 08:40:49.411094 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6c6f84469f-k22bc"] Dec 17 08:40:49 crc kubenswrapper[4966]: I1217 08:40:49.612175 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:40:49 crc kubenswrapper[4966]: W1217 08:40:49.660198 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67cdcfd5_3af2_475d_bb79_f91357e1b0bd.slice/crio-97f1ac08aa28a49a45d392c10460025c7551080a4bbafb8d130bd21a3ce31d33 WatchSource:0}: Error finding container 97f1ac08aa28a49a45d392c10460025c7551080a4bbafb8d130bd21a3ce31d33: Status 404 returned error can't find the container with id 97f1ac08aa28a49a45d392c10460025c7551080a4bbafb8d130bd21a3ce31d33 Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.110750 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5679cd5d4c-bstjs" event={"ID":"0333b211-1df2-45b3-9f53-bfcc921f98cc","Type":"ContainerStarted","Data":"187259dd2caa1fe0d1df1b18dc205d58df63e73d68956098673e5584fd77772c"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.146368 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerStarted","Data":"97f1ac08aa28a49a45d392c10460025c7551080a4bbafb8d130bd21a3ce31d33"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.160828 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f66c49f7b-4992d" event={"ID":"707f4037-3896-482d-85a8-3b0fc081045d","Type":"ContainerStarted","Data":"cf1d6a2e774f8312f90fce4e2cf6ab859839ecd1cefc7a81f2fcb8bfc6ae482e"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.162425 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" event={"ID":"6c618545-5898-4e00-9ed3-5f34f2ad77c8","Type":"ContainerStarted","Data":"72f776d3e93779e4cbc17a76db405133174c02f728a50829e3bb4b99e93b517b"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.173321 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" event={"ID":"7fb54742-7249-4b48-a918-1bfab858a763","Type":"ContainerStarted","Data":"a4a4f9b441601fb5b2ec86eb01abdb18b1fb6827411014b5de9031bbe0515bd9"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.173371 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" event={"ID":"7fb54742-7249-4b48-a918-1bfab858a763","Type":"ContainerStarted","Data":"077204e63e9658c00f670a0e802a7378f50202342a418d8a943cffcd1382b609"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.174520 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" podStartSLOduration=9.174503047 podStartE2EDuration="9.174503047s" podCreationTimestamp="2025-12-17 08:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:50.174359093 +0000 UTC m=+1185.719429035" watchObservedRunningTime="2025-12-17 08:40:50.174503047 +0000 UTC m=+1185.719572989" Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.187130 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-68976d496b-zns8c" event={"ID":"f1e350c4-e807-4015-8398-42bad206482f","Type":"ContainerStarted","Data":"3289011c10de198012267280261cb3953329945f3cd0d5cd2f7a06412dcd6f17"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.189090 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6c6f84469f-k22bc" event={"ID":"bc264f63-3c8c-47d7-851b-7fcbcac72b5e","Type":"ContainerStarted","Data":"1c3a3ac1c6a3509c21756c6beaeeb0bf7eed1cbbd2ffd7f2913c8c3138ed3498"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.221649 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-spxpd" event={"ID":"2b302878-13e7-4967-aaf1-58292a7237e5","Type":"ContainerStarted","Data":"abe99bca543fdf2cf906ae9e9294ec48c8366d00058c0213a01458e6fb3c7664"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.221709 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-spxpd" event={"ID":"2b302878-13e7-4967-aaf1-58292a7237e5","Type":"ContainerStarted","Data":"b91a247e8f487d8e40d8b5f034ac529e52546be0f1474e1cdc6d7e2f67f2ac0f"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.242549 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" event={"ID":"fed2ac76-ebb0-481f-82e1-55397f658c99","Type":"ContainerStarted","Data":"af47b553eb89f736be25a50d8ab3e9a39064e822b0f742f58fa47098644aa92f"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.242594 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" event={"ID":"fed2ac76-ebb0-481f-82e1-55397f658c99","Type":"ContainerStarted","Data":"fde5a476be2bfcf90c336781c743ab7152addfe3156a7d2600ddbaf2a7733020"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.263197 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d41-account-create-update-x4fn4" event={"ID":"ce426531-9f5f-4c9d-8bea-e22d666fe95d","Type":"ContainerStarted","Data":"5ba9b49c17b666885899786efb3d50937ceb85900c2a06b6198056f893f3000e"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.263239 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d41-account-create-update-x4fn4" event={"ID":"ce426531-9f5f-4c9d-8bea-e22d666fe95d","Type":"ContainerStarted","Data":"5bde4481fd86ebf98bb30fc7c150b3b99ef8994f101dd468893e6a9d2cb91a19"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.265658 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-spxpd" podStartSLOduration=9.265617574 podStartE2EDuration="9.265617574s" podCreationTimestamp="2025-12-17 08:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:50.236537238 +0000 UTC m=+1185.781607200" watchObservedRunningTime="2025-12-17 08:40:50.265617574 +0000 UTC m=+1185.810687516" Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.290101 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" podStartSLOduration=9.290079845 podStartE2EDuration="9.290079845s" podCreationTimestamp="2025-12-17 08:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:50.263025304 +0000 UTC m=+1185.808095246" watchObservedRunningTime="2025-12-17 08:40:50.290079845 +0000 UTC m=+1185.835149797" Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.293226 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6cb857dc54-5f9tj" event={"ID":"1a20b1b5-aab0-4f67-ad55-023e9c908dd6","Type":"ContainerStarted","Data":"2ce1863e1ab4974e40e63d0306743c5e42c11b2a7702daeb106ee683ebf4d9d6"} Dec 17 08:40:50 crc kubenswrapper[4966]: I1217 08:40:50.307213 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-7d41-account-create-update-x4fn4" podStartSLOduration=9.307188204 podStartE2EDuration="9.307188204s" podCreationTimestamp="2025-12-17 08:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:50.283659299 +0000 UTC m=+1185.828729241" watchObservedRunningTime="2025-12-17 08:40:50.307188204 +0000 UTC m=+1185.852258146" Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.307347 4966 generic.go:334] "Generic (PLEG): container finished" podID="fed2ac76-ebb0-481f-82e1-55397f658c99" containerID="af47b553eb89f736be25a50d8ab3e9a39064e822b0f742f58fa47098644aa92f" exitCode=0 Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.307747 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" event={"ID":"fed2ac76-ebb0-481f-82e1-55397f658c99","Type":"ContainerDied","Data":"af47b553eb89f736be25a50d8ab3e9a39064e822b0f742f58fa47098644aa92f"} Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.315118 4966 generic.go:334] "Generic (PLEG): container finished" podID="ce426531-9f5f-4c9d-8bea-e22d666fe95d" containerID="5ba9b49c17b666885899786efb3d50937ceb85900c2a06b6198056f893f3000e" exitCode=0 Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.315185 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d41-account-create-update-x4fn4" event={"ID":"ce426531-9f5f-4c9d-8bea-e22d666fe95d","Type":"ContainerDied","Data":"5ba9b49c17b666885899786efb3d50937ceb85900c2a06b6198056f893f3000e"} Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.319776 4966 generic.go:334] "Generic (PLEG): container finished" podID="7fb54742-7249-4b48-a918-1bfab858a763" containerID="a4a4f9b441601fb5b2ec86eb01abdb18b1fb6827411014b5de9031bbe0515bd9" exitCode=0 Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.320022 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" event={"ID":"7fb54742-7249-4b48-a918-1bfab858a763","Type":"ContainerDied","Data":"a4a4f9b441601fb5b2ec86eb01abdb18b1fb6827411014b5de9031bbe0515bd9"} Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.325608 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5679cd5d4c-bstjs" event={"ID":"0333b211-1df2-45b3-9f53-bfcc921f98cc","Type":"ContainerStarted","Data":"09dd15dad694df8ee002a352bdb79dcdac1dba068511dbdeb973ce8ee5c50eba"} Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.328953 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerStarted","Data":"35d10ea52b1c854bf8b185fb2c6a83221681162903ab3b1f481412088424d8da"} Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.330139 4966 generic.go:334] "Generic (PLEG): container finished" podID="c9ea912d-6078-4ab1-8188-f1e2927eb12a" containerID="c03aac27db194fa66d8f77291e0f1b16dda26943f297720177bd3bf48ed980a5" exitCode=0 Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.330183 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" event={"ID":"c9ea912d-6078-4ab1-8188-f1e2927eb12a","Type":"ContainerDied","Data":"c03aac27db194fa66d8f77291e0f1b16dda26943f297720177bd3bf48ed980a5"} Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.331236 4966 generic.go:334] "Generic (PLEG): container finished" podID="2b302878-13e7-4967-aaf1-58292a7237e5" containerID="abe99bca543fdf2cf906ae9e9294ec48c8366d00058c0213a01458e6fb3c7664" exitCode=0 Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.331283 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-spxpd" event={"ID":"2b302878-13e7-4967-aaf1-58292a7237e5","Type":"ContainerDied","Data":"abe99bca543fdf2cf906ae9e9294ec48c8366d00058c0213a01458e6fb3c7664"} Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.345886 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f66c49f7b-4992d" event={"ID":"707f4037-3896-482d-85a8-3b0fc081045d","Type":"ContainerStarted","Data":"21e32b078c7b786d0fc87260d09c155e4d48123dca8bfb8a12ec99dc4b86b57a"} Dec 17 08:40:51 crc kubenswrapper[4966]: I1217 08:40:51.417345 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7f66c49f7b-4992d" podStartSLOduration=8.417324273 podStartE2EDuration="8.417324273s" podCreationTimestamp="2025-12-17 08:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:51.394804175 +0000 UTC m=+1186.939874127" watchObservedRunningTime="2025-12-17 08:40:51.417324273 +0000 UTC m=+1186.962394215" Dec 17 08:40:52 crc kubenswrapper[4966]: I1217 08:40:52.379349 4966 generic.go:334] "Generic (PLEG): container finished" podID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerID="c24b7f60ad38b0edeedf9daa35fb224111274921103a9efaaa7ef438157c889c" exitCode=137 Dec 17 08:40:52 crc kubenswrapper[4966]: I1217 08:40:52.379425 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fbfdd886-69frk" event={"ID":"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017","Type":"ContainerDied","Data":"c24b7f60ad38b0edeedf9daa35fb224111274921103a9efaaa7ef438157c889c"} Dec 17 08:40:52 crc kubenswrapper[4966]: I1217 08:40:52.380133 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.428477 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" event={"ID":"fed2ac76-ebb0-481f-82e1-55397f658c99","Type":"ContainerDied","Data":"fde5a476be2bfcf90c336781c743ab7152addfe3156a7d2600ddbaf2a7733020"} Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.428898 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fde5a476be2bfcf90c336781c743ab7152addfe3156a7d2600ddbaf2a7733020" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.434042 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d41-account-create-update-x4fn4" event={"ID":"ce426531-9f5f-4c9d-8bea-e22d666fe95d","Type":"ContainerDied","Data":"5bde4481fd86ebf98bb30fc7c150b3b99ef8994f101dd468893e6a9d2cb91a19"} Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.434064 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bde4481fd86ebf98bb30fc7c150b3b99ef8994f101dd468893e6a9d2cb91a19" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.450830 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xlrkm" event={"ID":"0644984a-4f00-4b7a-9a05-d9e43097a7c4","Type":"ContainerDied","Data":"4ed04c2942e721e99e697deb49ab147665bc91012a929f3b49cca66e560f18f1"} Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.450884 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ed04c2942e721e99e697deb49ab147665bc91012a929f3b49cca66e560f18f1" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.455172 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.455897 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" event={"ID":"c9ea912d-6078-4ab1-8188-f1e2927eb12a","Type":"ContainerDied","Data":"2d391d3f287f5b44e7cf194a14bc5854f51435b9a721415cf886fbdcae260051"} Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.455921 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d391d3f287f5b44e7cf194a14bc5854f51435b9a721415cf886fbdcae260051" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.463383 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-spxpd" event={"ID":"2b302878-13e7-4967-aaf1-58292a7237e5","Type":"ContainerDied","Data":"b91a247e8f487d8e40d8b5f034ac529e52546be0f1474e1cdc6d7e2f67f2ac0f"} Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.463406 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b91a247e8f487d8e40d8b5f034ac529e52546be0f1474e1cdc6d7e2f67f2ac0f" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.466297 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.469462 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.511737 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c4llv" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.517230 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c4llv" event={"ID":"6b05f9d6-237c-4cb0-91b9-505715528c7d","Type":"ContainerDied","Data":"3532ff265744c4eec436627e330a9c4913e355a8d467950afd2e251654b91608"} Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.517259 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3532ff265744c4eec436627e330a9c4913e355a8d467950afd2e251654b91608" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.525269 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.541238 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.542003 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b05f9d6-237c-4cb0-91b9-505715528c7d-operator-scripts\") pod \"6b05f9d6-237c-4cb0-91b9-505715528c7d\" (UID: \"6b05f9d6-237c-4cb0-91b9-505715528c7d\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.542156 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b302878-13e7-4967-aaf1-58292a7237e5-operator-scripts\") pod \"2b302878-13e7-4967-aaf1-58292a7237e5\" (UID: \"2b302878-13e7-4967-aaf1-58292a7237e5\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.542203 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ckl4\" (UniqueName: \"kubernetes.io/projected/6b05f9d6-237c-4cb0-91b9-505715528c7d-kube-api-access-2ckl4\") pod \"6b05f9d6-237c-4cb0-91b9-505715528c7d\" (UID: \"6b05f9d6-237c-4cb0-91b9-505715528c7d\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.542240 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr7sv\" (UniqueName: \"kubernetes.io/projected/2b302878-13e7-4967-aaf1-58292a7237e5-kube-api-access-wr7sv\") pod \"2b302878-13e7-4967-aaf1-58292a7237e5\" (UID: \"2b302878-13e7-4967-aaf1-58292a7237e5\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.543315 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b05f9d6-237c-4cb0-91b9-505715528c7d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6b05f9d6-237c-4cb0-91b9-505715528c7d" (UID: "6b05f9d6-237c-4cb0-91b9-505715528c7d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.543632 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b302878-13e7-4967-aaf1-58292a7237e5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2b302878-13e7-4967-aaf1-58292a7237e5" (UID: "2b302878-13e7-4967-aaf1-58292a7237e5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.548037 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b05f9d6-237c-4cb0-91b9-505715528c7d-kube-api-access-2ckl4" (OuterVolumeSpecName: "kube-api-access-2ckl4") pod "6b05f9d6-237c-4cb0-91b9-505715528c7d" (UID: "6b05f9d6-237c-4cb0-91b9-505715528c7d"). InnerVolumeSpecName "kube-api-access-2ckl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.549629 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b302878-13e7-4967-aaf1-58292a7237e5-kube-api-access-wr7sv" (OuterVolumeSpecName: "kube-api-access-wr7sv") pod "2b302878-13e7-4967-aaf1-58292a7237e5" (UID: "2b302878-13e7-4967-aaf1-58292a7237e5"). InnerVolumeSpecName "kube-api-access-wr7sv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.558863 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.647461 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdhrw\" (UniqueName: \"kubernetes.io/projected/fed2ac76-ebb0-481f-82e1-55397f658c99-kube-api-access-tdhrw\") pod \"fed2ac76-ebb0-481f-82e1-55397f658c99\" (UID: \"fed2ac76-ebb0-481f-82e1-55397f658c99\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.647502 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0644984a-4f00-4b7a-9a05-d9e43097a7c4-operator-scripts\") pod \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\" (UID: \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.647563 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed2ac76-ebb0-481f-82e1-55397f658c99-operator-scripts\") pod \"fed2ac76-ebb0-481f-82e1-55397f658c99\" (UID: \"fed2ac76-ebb0-481f-82e1-55397f658c99\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.647663 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55hq4\" (UniqueName: \"kubernetes.io/projected/0644984a-4f00-4b7a-9a05-d9e43097a7c4-kube-api-access-55hq4\") pod \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\" (UID: \"0644984a-4f00-4b7a-9a05-d9e43097a7c4\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.647698 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rn2n\" (UniqueName: \"kubernetes.io/projected/c9ea912d-6078-4ab1-8188-f1e2927eb12a-kube-api-access-2rn2n\") pod \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\" (UID: \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.647790 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9ea912d-6078-4ab1-8188-f1e2927eb12a-operator-scripts\") pod \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\" (UID: \"c9ea912d-6078-4ab1-8188-f1e2927eb12a\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.647815 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s9vw\" (UniqueName: \"kubernetes.io/projected/ce426531-9f5f-4c9d-8bea-e22d666fe95d-kube-api-access-7s9vw\") pod \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\" (UID: \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.647834 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce426531-9f5f-4c9d-8bea-e22d666fe95d-operator-scripts\") pod \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\" (UID: \"ce426531-9f5f-4c9d-8bea-e22d666fe95d\") " Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.648336 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b302878-13e7-4967-aaf1-58292a7237e5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.648350 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ckl4\" (UniqueName: \"kubernetes.io/projected/6b05f9d6-237c-4cb0-91b9-505715528c7d-kube-api-access-2ckl4\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.648359 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr7sv\" (UniqueName: \"kubernetes.io/projected/2b302878-13e7-4967-aaf1-58292a7237e5-kube-api-access-wr7sv\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.648368 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b05f9d6-237c-4cb0-91b9-505715528c7d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.653424 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9ea912d-6078-4ab1-8188-f1e2927eb12a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c9ea912d-6078-4ab1-8188-f1e2927eb12a" (UID: "c9ea912d-6078-4ab1-8188-f1e2927eb12a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.653814 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0644984a-4f00-4b7a-9a05-d9e43097a7c4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0644984a-4f00-4b7a-9a05-d9e43097a7c4" (UID: "0644984a-4f00-4b7a-9a05-d9e43097a7c4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.654164 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fed2ac76-ebb0-481f-82e1-55397f658c99-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fed2ac76-ebb0-481f-82e1-55397f658c99" (UID: "fed2ac76-ebb0-481f-82e1-55397f658c99"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.662611 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce426531-9f5f-4c9d-8bea-e22d666fe95d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ce426531-9f5f-4c9d-8bea-e22d666fe95d" (UID: "ce426531-9f5f-4c9d-8bea-e22d666fe95d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.664156 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce426531-9f5f-4c9d-8bea-e22d666fe95d-kube-api-access-7s9vw" (OuterVolumeSpecName: "kube-api-access-7s9vw") pod "ce426531-9f5f-4c9d-8bea-e22d666fe95d" (UID: "ce426531-9f5f-4c9d-8bea-e22d666fe95d"). InnerVolumeSpecName "kube-api-access-7s9vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.664497 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fed2ac76-ebb0-481f-82e1-55397f658c99-kube-api-access-tdhrw" (OuterVolumeSpecName: "kube-api-access-tdhrw") pod "fed2ac76-ebb0-481f-82e1-55397f658c99" (UID: "fed2ac76-ebb0-481f-82e1-55397f658c99"). InnerVolumeSpecName "kube-api-access-tdhrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.670057 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0644984a-4f00-4b7a-9a05-d9e43097a7c4-kube-api-access-55hq4" (OuterVolumeSpecName: "kube-api-access-55hq4") pod "0644984a-4f00-4b7a-9a05-d9e43097a7c4" (UID: "0644984a-4f00-4b7a-9a05-d9e43097a7c4"). InnerVolumeSpecName "kube-api-access-55hq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.670205 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9ea912d-6078-4ab1-8188-f1e2927eb12a-kube-api-access-2rn2n" (OuterVolumeSpecName: "kube-api-access-2rn2n") pod "c9ea912d-6078-4ab1-8188-f1e2927eb12a" (UID: "c9ea912d-6078-4ab1-8188-f1e2927eb12a"). InnerVolumeSpecName "kube-api-access-2rn2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.750408 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55hq4\" (UniqueName: \"kubernetes.io/projected/0644984a-4f00-4b7a-9a05-d9e43097a7c4-kube-api-access-55hq4\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.750438 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rn2n\" (UniqueName: \"kubernetes.io/projected/c9ea912d-6078-4ab1-8188-f1e2927eb12a-kube-api-access-2rn2n\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.750450 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c9ea912d-6078-4ab1-8188-f1e2927eb12a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.750459 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s9vw\" (UniqueName: \"kubernetes.io/projected/ce426531-9f5f-4c9d-8bea-e22d666fe95d-kube-api-access-7s9vw\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.750472 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce426531-9f5f-4c9d-8bea-e22d666fe95d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.750484 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdhrw\" (UniqueName: \"kubernetes.io/projected/fed2ac76-ebb0-481f-82e1-55397f658c99-kube-api-access-tdhrw\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.750497 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0644984a-4f00-4b7a-9a05-d9e43097a7c4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:53 crc kubenswrapper[4966]: I1217 08:40:53.750508 4966 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fed2ac76-ebb0-481f-82e1-55397f658c99-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.003682 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.161895 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-config-data\") pod \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.161953 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xvb9\" (UniqueName: \"kubernetes.io/projected/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-kube-api-access-4xvb9\") pod \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.162006 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-scripts\") pod \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.162748 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-tls-certs\") pod \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.162806 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-combined-ca-bundle\") pod \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.162821 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-logs\") pod \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.162847 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-secret-key\") pod \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\" (UID: \"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017\") " Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.164856 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-logs" (OuterVolumeSpecName: "logs") pod "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" (UID: "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.178109 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-kube-api-access-4xvb9" (OuterVolumeSpecName: "kube-api-access-4xvb9") pod "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" (UID: "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017"). InnerVolumeSpecName "kube-api-access-4xvb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.183729 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" (UID: "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.265600 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.265641 4966 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.265650 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xvb9\" (UniqueName: \"kubernetes.io/projected/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-kube-api-access-4xvb9\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.327575 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" (UID: "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.339309 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-scripts" (OuterVolumeSpecName: "scripts") pod "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" (UID: "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.367784 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.368025 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.401530 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-config-data" (OuterVolumeSpecName: "config-data") pod "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" (UID: "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.406078 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" (UID: "04dbdf7d-ebdc-46b9-8b29-b6e427b5d017"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.470128 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.470385 4966 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.527768 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" event={"ID":"6c618545-5898-4e00-9ed3-5f34f2ad77c8","Type":"ContainerStarted","Data":"d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275"} Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.528102 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" podUID="6c618545-5898-4e00-9ed3-5f34f2ad77c8" containerName="heat-cfnapi" containerID="cri-o://d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275" gracePeriod=60 Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.528407 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.535801 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" event={"ID":"7fb54742-7249-4b48-a918-1bfab858a763","Type":"ContainerStarted","Data":"3ddc1883fa26c23d1faf2a962987f8ac9f9ac279fdf3b42dc24064cfccdcf8f5"} Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.535840 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.550244 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-68976d496b-zns8c" event={"ID":"f1e350c4-e807-4015-8398-42bad206482f","Type":"ContainerStarted","Data":"1b54c4c8f2a633b5bb9ce85714c60121e0a0bbfe5e92020c89de55031fcc1217"} Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.551335 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.559981 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-fbfdd886-69frk" event={"ID":"04dbdf7d-ebdc-46b9-8b29-b6e427b5d017","Type":"ContainerDied","Data":"7098111dbed79bb95427735d2299eb8c8e2ddb3e2ca55cf702267f4b20e66124"} Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.560205 4966 scope.go:117] "RemoveContainer" containerID="ff860a43024f6954192167e387c152dc2b8c1a2ad2cbeaee0c6229e8eb0dd637" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.560323 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-fbfdd886-69frk" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.579866 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-e4a6-account-create-update-q8vbr" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.580169 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-spxpd" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.581379 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5679cd5d4c-bstjs" event={"ID":"0333b211-1df2-45b3-9f53-bfcc921f98cc","Type":"ContainerStarted","Data":"a6c2a47f4df817319728c7cdc1dd762077bfe1d62fccd0c314f3d9b527454521"} Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.581508 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.581594 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.581715 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-67a5-account-create-update-qdhbr" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.589087 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d41-account-create-update-x4fn4" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.592132 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xlrkm" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.611159 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" podStartSLOduration=14.416649884 podStartE2EDuration="18.611138904s" podCreationTimestamp="2025-12-17 08:40:36 +0000 UTC" firstStartedPulling="2025-12-17 08:40:49.385767538 +0000 UTC m=+1184.930837480" lastFinishedPulling="2025-12-17 08:40:53.580256558 +0000 UTC m=+1189.125326500" observedRunningTime="2025-12-17 08:40:54.563405166 +0000 UTC m=+1190.108475108" watchObservedRunningTime="2025-12-17 08:40:54.611138904 +0000 UTC m=+1190.156208846" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.634471 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" podStartSLOduration=18.634455043 podStartE2EDuration="18.634455043s" podCreationTimestamp="2025-12-17 08:40:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:54.612624885 +0000 UTC m=+1190.157694857" watchObservedRunningTime="2025-12-17 08:40:54.634455043 +0000 UTC m=+1190.179524985" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.709565 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-68976d496b-zns8c" podStartSLOduration=6.191797261 podStartE2EDuration="10.70951703s" podCreationTimestamp="2025-12-17 08:40:44 +0000 UTC" firstStartedPulling="2025-12-17 08:40:49.041385839 +0000 UTC m=+1184.586455781" lastFinishedPulling="2025-12-17 08:40:53.559105608 +0000 UTC m=+1189.104175550" observedRunningTime="2025-12-17 08:40:54.670078929 +0000 UTC m=+1190.215148881" watchObservedRunningTime="2025-12-17 08:40:54.70951703 +0000 UTC m=+1190.254586972" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.724136 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5679cd5d4c-bstjs" podStartSLOduration=13.724115051 podStartE2EDuration="13.724115051s" podCreationTimestamp="2025-12-17 08:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:40:54.694292383 +0000 UTC m=+1190.239362325" watchObservedRunningTime="2025-12-17 08:40:54.724115051 +0000 UTC m=+1190.269184993" Dec 17 08:40:54 crc kubenswrapper[4966]: I1217 08:40:54.820029 4966 scope.go:117] "RemoveContainer" containerID="c24b7f60ad38b0edeedf9daa35fb224111274921103a9efaaa7ef438157c889c" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.090115 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-fbfdd886-69frk"] Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.098606 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-fbfdd886-69frk"] Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.587993 4966 generic.go:334] "Generic (PLEG): container finished" podID="f1e350c4-e807-4015-8398-42bad206482f" containerID="1b54c4c8f2a633b5bb9ce85714c60121e0a0bbfe5e92020c89de55031fcc1217" exitCode=1 Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.588053 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-68976d496b-zns8c" event={"ID":"f1e350c4-e807-4015-8398-42bad206482f","Type":"ContainerDied","Data":"1b54c4c8f2a633b5bb9ce85714c60121e0a0bbfe5e92020c89de55031fcc1217"} Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.588663 4966 scope.go:117] "RemoveContainer" containerID="1b54c4c8f2a633b5bb9ce85714c60121e0a0bbfe5e92020c89de55031fcc1217" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.593299 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerStarted","Data":"4f790cb8cb7a6f66eef74cdfe3c1adb0471632552a25bc333210d5ec2718315f"} Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.595397 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6c6f84469f-k22bc" event={"ID":"bc264f63-3c8c-47d7-851b-7fcbcac72b5e","Type":"ContainerStarted","Data":"6ca98b4c41d2890ea4d566533a1323f97e60428e0c792ee79f3a34356704e9b6"} Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.595848 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.606050 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6cb857dc54-5f9tj" event={"ID":"1a20b1b5-aab0-4f67-ad55-023e9c908dd6","Type":"ContainerStarted","Data":"941c9c4e0774962dd8388032519d216bbd80c191fd78bc651231b9d9105720bd"} Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.606887 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.620667 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" event={"ID":"cdfc5775-0ec0-433c-9774-0e6a56273b8e","Type":"ContainerStarted","Data":"fd99b8f9f1db1e8f333cb8811cf6e3e40b9e62c4ffd884c0e8a04ac892d209b6"} Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.620985 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.651495 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6c6f84469f-k22bc" podStartSLOduration=4.483477757 podStartE2EDuration="8.651473779s" podCreationTimestamp="2025-12-17 08:40:47 +0000 UTC" firstStartedPulling="2025-12-17 08:40:49.413060467 +0000 UTC m=+1184.958130399" lastFinishedPulling="2025-12-17 08:40:53.581056479 +0000 UTC m=+1189.126126421" observedRunningTime="2025-12-17 08:40:55.635484271 +0000 UTC m=+1191.180554223" watchObservedRunningTime="2025-12-17 08:40:55.651473779 +0000 UTC m=+1191.196543721" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.657420 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-6c76d785bf-vjmcl" podUID="0b528e6e-b726-4895-910c-8544dcf26417" containerName="heat-api" containerID="cri-o://e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c" gracePeriod=60 Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.657740 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6c76d785bf-vjmcl" event={"ID":"0b528e6e-b726-4895-910c-8544dcf26417","Type":"ContainerStarted","Data":"e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c"} Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.658471 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.673060 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6cb857dc54-5f9tj" podStartSLOduration=4.569539475 podStartE2EDuration="8.673040981s" podCreationTimestamp="2025-12-17 08:40:47 +0000 UTC" firstStartedPulling="2025-12-17 08:40:49.389722947 +0000 UTC m=+1184.934792889" lastFinishedPulling="2025-12-17 08:40:53.493224453 +0000 UTC m=+1189.038294395" observedRunningTime="2025-12-17 08:40:55.660623731 +0000 UTC m=+1191.205693673" watchObservedRunningTime="2025-12-17 08:40:55.673040981 +0000 UTC m=+1191.218110923" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.681292 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5679cd5d4c-bstjs" podUID="0333b211-1df2-45b3-9f53-bfcc921f98cc" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.693465 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6c76d785bf-vjmcl" podStartSLOduration=13.731143445 podStartE2EDuration="19.69344535s" podCreationTimestamp="2025-12-17 08:40:36 +0000 UTC" firstStartedPulling="2025-12-17 08:40:47.611899887 +0000 UTC m=+1183.156969829" lastFinishedPulling="2025-12-17 08:40:53.574201792 +0000 UTC m=+1189.119271734" observedRunningTime="2025-12-17 08:40:55.690330455 +0000 UTC m=+1191.235400407" watchObservedRunningTime="2025-12-17 08:40:55.69344535 +0000 UTC m=+1191.238515292" Dec 17 08:40:55 crc kubenswrapper[4966]: I1217 08:40:55.740383 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" podStartSLOduration=6.168304407 podStartE2EDuration="11.740359696s" podCreationTimestamp="2025-12-17 08:40:44 +0000 UTC" firstStartedPulling="2025-12-17 08:40:47.921140193 +0000 UTC m=+1183.466210135" lastFinishedPulling="2025-12-17 08:40:53.493195482 +0000 UTC m=+1189.038265424" observedRunningTime="2025-12-17 08:40:55.718955869 +0000 UTC m=+1191.264025811" watchObservedRunningTime="2025-12-17 08:40:55.740359696 +0000 UTC m=+1191.285429638" Dec 17 08:40:56 crc kubenswrapper[4966]: I1217 08:40:56.711814 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-68976d496b-zns8c" event={"ID":"f1e350c4-e807-4015-8398-42bad206482f","Type":"ContainerStarted","Data":"aa972212fd850218498cff0b14105afa9d738631c7a214f0dee83d10477ae72e"} Dec 17 08:40:56 crc kubenswrapper[4966]: I1217 08:40:56.715027 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:56 crc kubenswrapper[4966]: I1217 08:40:56.732977 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerStarted","Data":"90efbdbe12e50f6e45148a17f90d31339da3da4aeddd5b010295c46f04149b98"} Dec 17 08:40:56 crc kubenswrapper[4966]: I1217 08:40:56.734541 4966 generic.go:334] "Generic (PLEG): container finished" podID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" containerID="fd99b8f9f1db1e8f333cb8811cf6e3e40b9e62c4ffd884c0e8a04ac892d209b6" exitCode=1 Dec 17 08:40:56 crc kubenswrapper[4966]: I1217 08:40:56.735923 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" event={"ID":"cdfc5775-0ec0-433c-9774-0e6a56273b8e","Type":"ContainerDied","Data":"fd99b8f9f1db1e8f333cb8811cf6e3e40b9e62c4ffd884c0e8a04ac892d209b6"} Dec 17 08:40:56 crc kubenswrapper[4966]: I1217 08:40:56.760664 4966 scope.go:117] "RemoveContainer" containerID="fd99b8f9f1db1e8f333cb8811cf6e3e40b9e62c4ffd884c0e8a04ac892d209b6" Dec 17 08:40:56 crc kubenswrapper[4966]: I1217 08:40:56.912315 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" path="/var/lib/kubelet/pods/04dbdf7d-ebdc-46b9-8b29-b6e427b5d017/volumes" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.222094 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzwzg"] Dec 17 08:40:57 crc kubenswrapper[4966]: E1217 08:40:57.222733 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce426531-9f5f-4c9d-8bea-e22d666fe95d" containerName="mariadb-account-create-update" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.222749 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce426531-9f5f-4c9d-8bea-e22d666fe95d" containerName="mariadb-account-create-update" Dec 17 08:40:57 crc kubenswrapper[4966]: E1217 08:40:57.222771 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b302878-13e7-4967-aaf1-58292a7237e5" containerName="mariadb-database-create" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.222778 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b302878-13e7-4967-aaf1-58292a7237e5" containerName="mariadb-database-create" Dec 17 08:40:57 crc kubenswrapper[4966]: E1217 08:40:57.222785 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b05f9d6-237c-4cb0-91b9-505715528c7d" containerName="mariadb-database-create" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.222792 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b05f9d6-237c-4cb0-91b9-505715528c7d" containerName="mariadb-database-create" Dec 17 08:40:57 crc kubenswrapper[4966]: E1217 08:40:57.222810 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed2ac76-ebb0-481f-82e1-55397f658c99" containerName="mariadb-account-create-update" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.222818 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed2ac76-ebb0-481f-82e1-55397f658c99" containerName="mariadb-account-create-update" Dec 17 08:40:57 crc kubenswrapper[4966]: E1217 08:40:57.222826 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon-log" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.222832 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon-log" Dec 17 08:40:57 crc kubenswrapper[4966]: E1217 08:40:57.222843 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0644984a-4f00-4b7a-9a05-d9e43097a7c4" containerName="mariadb-database-create" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.222849 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0644984a-4f00-4b7a-9a05-d9e43097a7c4" containerName="mariadb-database-create" Dec 17 08:40:57 crc kubenswrapper[4966]: E1217 08:40:57.222857 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ea912d-6078-4ab1-8188-f1e2927eb12a" containerName="mariadb-account-create-update" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.222862 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ea912d-6078-4ab1-8188-f1e2927eb12a" containerName="mariadb-account-create-update" Dec 17 08:40:57 crc kubenswrapper[4966]: E1217 08:40:57.222893 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.222899 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.223066 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b302878-13e7-4967-aaf1-58292a7237e5" containerName="mariadb-database-create" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.223086 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon-log" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.223095 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9ea912d-6078-4ab1-8188-f1e2927eb12a" containerName="mariadb-account-create-update" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.223105 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b05f9d6-237c-4cb0-91b9-505715528c7d" containerName="mariadb-database-create" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.223112 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce426531-9f5f-4c9d-8bea-e22d666fe95d" containerName="mariadb-account-create-update" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.223120 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="04dbdf7d-ebdc-46b9-8b29-b6e427b5d017" containerName="horizon" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.223136 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="fed2ac76-ebb0-481f-82e1-55397f658c99" containerName="mariadb-account-create-update" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.223147 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0644984a-4f00-4b7a-9a05-d9e43097a7c4" containerName="mariadb-database-create" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.223760 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.245968 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzwzg"] Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.253590 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s6hbj" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.262734 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.262906 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.412300 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-config-data\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.412726 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9j4c\" (UniqueName: \"kubernetes.io/projected/02f6a226-d49b-4726-8116-4b9ab2837338-kube-api-access-n9j4c\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.412929 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-scripts\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.413175 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.459671 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5679cd5d4c-bstjs" podUID="0333b211-1df2-45b3-9f53-bfcc921f98cc" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.460580 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-5679cd5d4c-bstjs" podUID="0333b211-1df2-45b3-9f53-bfcc921f98cc" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.514381 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9j4c\" (UniqueName: \"kubernetes.io/projected/02f6a226-d49b-4726-8116-4b9ab2837338-kube-api-access-n9j4c\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.514791 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-scripts\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.515022 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.515132 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-config-data\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.520198 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-config-data\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.523536 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.530310 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-scripts\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.555222 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9j4c\" (UniqueName: \"kubernetes.io/projected/02f6a226-d49b-4726-8116-4b9ab2837338-kube-api-access-n9j4c\") pod \"nova-cell0-conductor-db-sync-hzwzg\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.743780 4966 generic.go:334] "Generic (PLEG): container finished" podID="f1e350c4-e807-4015-8398-42bad206482f" containerID="aa972212fd850218498cff0b14105afa9d738631c7a214f0dee83d10477ae72e" exitCode=1 Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.743833 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-68976d496b-zns8c" event={"ID":"f1e350c4-e807-4015-8398-42bad206482f","Type":"ContainerDied","Data":"aa972212fd850218498cff0b14105afa9d738631c7a214f0dee83d10477ae72e"} Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.743864 4966 scope.go:117] "RemoveContainer" containerID="1b54c4c8f2a633b5bb9ce85714c60121e0a0bbfe5e92020c89de55031fcc1217" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.744454 4966 scope.go:117] "RemoveContainer" containerID="aa972212fd850218498cff0b14105afa9d738631c7a214f0dee83d10477ae72e" Dec 17 08:40:57 crc kubenswrapper[4966]: E1217 08:40:57.744730 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-68976d496b-zns8c_openstack(f1e350c4-e807-4015-8398-42bad206482f)\"" pod="openstack/heat-api-68976d496b-zns8c" podUID="f1e350c4-e807-4015-8398-42bad206482f" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.749673 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" event={"ID":"cdfc5775-0ec0-433c-9774-0e6a56273b8e","Type":"ContainerStarted","Data":"001b9b2ac8e52ce9138da62c7bd2d25d5c388b4978e9c26ab219cd343691f642"} Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.750244 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:57 crc kubenswrapper[4966]: I1217 08:40:57.844563 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.497690 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzwzg"] Dec 17 08:40:58 crc kubenswrapper[4966]: W1217 08:40:58.519081 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02f6a226_d49b_4726_8116_4b9ab2837338.slice/crio-930ffc4b129f2613ecb947f56b3691c7fb52bcd33d1991639d4acb6994200418 WatchSource:0}: Error finding container 930ffc4b129f2613ecb947f56b3691c7fb52bcd33d1991639d4acb6994200418: Status 404 returned error can't find the container with id 930ffc4b129f2613ecb947f56b3691c7fb52bcd33d1991639d4acb6994200418 Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.773561 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerStarted","Data":"fb31d48108cb02ca483a69aae35f6ba634c4e2999930e746821d372d6717a63d"} Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.773916 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.796353 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"220e53bd-2342-4bed-8e26-487e0859350b","Type":"ContainerStarted","Data":"fd277257dc5917e5618c66dfb7df21d7c4c2f4b92d19c4c3522e2b9e490f9f08"} Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.799626 4966 generic.go:334] "Generic (PLEG): container finished" podID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" containerID="001b9b2ac8e52ce9138da62c7bd2d25d5c388b4978e9c26ab219cd343691f642" exitCode=1 Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.799799 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" event={"ID":"cdfc5775-0ec0-433c-9774-0e6a56273b8e","Type":"ContainerDied","Data":"001b9b2ac8e52ce9138da62c7bd2d25d5c388b4978e9c26ab219cd343691f642"} Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.799913 4966 scope.go:117] "RemoveContainer" containerID="fd99b8f9f1db1e8f333cb8811cf6e3e40b9e62c4ffd884c0e8a04ac892d209b6" Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.801099 4966 scope.go:117] "RemoveContainer" containerID="001b9b2ac8e52ce9138da62c7bd2d25d5c388b4978e9c26ab219cd343691f642" Dec 17 08:40:58 crc kubenswrapper[4966]: E1217 08:40:58.801471 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-7668494dc4-2sqhx_openstack(cdfc5775-0ec0-433c-9774-0e6a56273b8e)\"" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.838496 4966 scope.go:117] "RemoveContainer" containerID="aa972212fd850218498cff0b14105afa9d738631c7a214f0dee83d10477ae72e" Dec 17 08:40:58 crc kubenswrapper[4966]: E1217 08:40:58.838766 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-68976d496b-zns8c_openstack(f1e350c4-e807-4015-8398-42bad206482f)\"" pod="openstack/heat-api-68976d496b-zns8c" podUID="f1e350c4-e807-4015-8398-42bad206482f" Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.850652 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.7677149869999997 podStartE2EDuration="11.850630717s" podCreationTimestamp="2025-12-17 08:40:47 +0000 UTC" firstStartedPulling="2025-12-17 08:40:49.672256041 +0000 UTC m=+1185.217325983" lastFinishedPulling="2025-12-17 08:40:57.755171771 +0000 UTC m=+1193.300241713" observedRunningTime="2025-12-17 08:40:58.838278969 +0000 UTC m=+1194.383348921" watchObservedRunningTime="2025-12-17 08:40:58.850630717 +0000 UTC m=+1194.395700659" Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.867971 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" event={"ID":"02f6a226-d49b-4726-8116-4b9ab2837338","Type":"ContainerStarted","Data":"930ffc4b129f2613ecb947f56b3691c7fb52bcd33d1991639d4acb6994200418"} Dec 17 08:40:58 crc kubenswrapper[4966]: I1217 08:40:58.890499 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.4215094329999998 podStartE2EDuration="29.890480549s" podCreationTimestamp="2025-12-17 08:40:29 +0000 UTC" firstStartedPulling="2025-12-17 08:40:30.471058502 +0000 UTC m=+1166.016128444" lastFinishedPulling="2025-12-17 08:40:57.940029618 +0000 UTC m=+1193.485099560" observedRunningTime="2025-12-17 08:40:58.878307426 +0000 UTC m=+1194.423377368" watchObservedRunningTime="2025-12-17 08:40:58.890480549 +0000 UTC m=+1194.435550491" Dec 17 08:40:59 crc kubenswrapper[4966]: I1217 08:40:59.384399 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:40:59 crc kubenswrapper[4966]: I1217 08:40:59.527217 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:40:59 crc kubenswrapper[4966]: I1217 08:40:59.885049 4966 scope.go:117] "RemoveContainer" containerID="aa972212fd850218498cff0b14105afa9d738631c7a214f0dee83d10477ae72e" Dec 17 08:40:59 crc kubenswrapper[4966]: E1217 08:40:59.885291 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-68976d496b-zns8c_openstack(f1e350c4-e807-4015-8398-42bad206482f)\"" pod="openstack/heat-api-68976d496b-zns8c" podUID="f1e350c4-e807-4015-8398-42bad206482f" Dec 17 08:40:59 crc kubenswrapper[4966]: I1217 08:40:59.886042 4966 scope.go:117] "RemoveContainer" containerID="001b9b2ac8e52ce9138da62c7bd2d25d5c388b4978e9c26ab219cd343691f642" Dec 17 08:40:59 crc kubenswrapper[4966]: E1217 08:40:59.886276 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-7668494dc4-2sqhx_openstack(cdfc5775-0ec0-433c-9774-0e6a56273b8e)\"" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" Dec 17 08:41:00 crc kubenswrapper[4966]: I1217 08:41:00.895076 4966 scope.go:117] "RemoveContainer" containerID="001b9b2ac8e52ce9138da62c7bd2d25d5c388b4978e9c26ab219cd343691f642" Dec 17 08:41:00 crc kubenswrapper[4966]: E1217 08:41:00.895497 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-7668494dc4-2sqhx_openstack(cdfc5775-0ec0-433c-9774-0e6a56273b8e)\"" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.056331 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.056609 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="ceilometer-central-agent" containerID="cri-o://35d10ea52b1c854bf8b185fb2c6a83221681162903ab3b1f481412088424d8da" gracePeriod=30 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.056755 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="proxy-httpd" containerID="cri-o://fb31d48108cb02ca483a69aae35f6ba634c4e2999930e746821d372d6717a63d" gracePeriod=30 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.056806 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="sg-core" containerID="cri-o://90efbdbe12e50f6e45148a17f90d31339da3da4aeddd5b010295c46f04149b98" gracePeriod=30 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.056848 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="ceilometer-notification-agent" containerID="cri-o://4f790cb8cb7a6f66eef74cdfe3c1adb0471632552a25bc333210d5ec2718315f" gracePeriod=30 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.452528 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.452851 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.455259 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5679cd5d4c-bstjs" Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.546258 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-798df684f5-9xzvn"] Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.546689 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerName="dnsmasq-dns" containerID="cri-o://37887270f964c260dd0744dae832d55f04a977e3e67d25a5dbe0c9863d3935dc" gracePeriod=10 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.939388 4966 generic.go:334] "Generic (PLEG): container finished" podID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerID="fb31d48108cb02ca483a69aae35f6ba634c4e2999930e746821d372d6717a63d" exitCode=0 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.939417 4966 generic.go:334] "Generic (PLEG): container finished" podID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerID="90efbdbe12e50f6e45148a17f90d31339da3da4aeddd5b010295c46f04149b98" exitCode=2 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.939426 4966 generic.go:334] "Generic (PLEG): container finished" podID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerID="4f790cb8cb7a6f66eef74cdfe3c1adb0471632552a25bc333210d5ec2718315f" exitCode=0 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.939432 4966 generic.go:334] "Generic (PLEG): container finished" podID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerID="35d10ea52b1c854bf8b185fb2c6a83221681162903ab3b1f481412088424d8da" exitCode=0 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.939489 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerDied","Data":"fb31d48108cb02ca483a69aae35f6ba634c4e2999930e746821d372d6717a63d"} Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.939547 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerDied","Data":"90efbdbe12e50f6e45148a17f90d31339da3da4aeddd5b010295c46f04149b98"} Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.939560 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerDied","Data":"4f790cb8cb7a6f66eef74cdfe3c1adb0471632552a25bc333210d5ec2718315f"} Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.939570 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerDied","Data":"35d10ea52b1c854bf8b185fb2c6a83221681162903ab3b1f481412088424d8da"} Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.944576 4966 generic.go:334] "Generic (PLEG): container finished" podID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerID="37887270f964c260dd0744dae832d55f04a977e3e67d25a5dbe0c9863d3935dc" exitCode=0 Dec 17 08:41:02 crc kubenswrapper[4966]: I1217 08:41:02.944649 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" event={"ID":"e3e02105-d411-4d80-9d37-ae44d4cd8a4c","Type":"ContainerDied","Data":"37887270f964c260dd0744dae832d55f04a977e3e67d25a5dbe0c9863d3935dc"} Dec 17 08:41:04 crc kubenswrapper[4966]: I1217 08:41:04.436942 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7f66c49f7b-4992d" Dec 17 08:41:04 crc kubenswrapper[4966]: I1217 08:41:04.489955 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-b67cfd874-bklxp"] Dec 17 08:41:04 crc kubenswrapper[4966]: I1217 08:41:04.490215 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-b67cfd874-bklxp" podUID="f33826e0-0269-4b99-9dce-fcaf62f67bed" containerName="heat-engine" containerID="cri-o://e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6" gracePeriod=60 Dec 17 08:41:04 crc kubenswrapper[4966]: E1217 08:41:04.508298 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 17 08:41:04 crc kubenswrapper[4966]: E1217 08:41:04.519308 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 17 08:41:04 crc kubenswrapper[4966]: E1217 08:41:04.544659 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 17 08:41:04 crc kubenswrapper[4966]: E1217 08:41:04.544745 4966 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-b67cfd874-bklxp" podUID="f33826e0-0269-4b99-9dce-fcaf62f67bed" containerName="heat-engine" Dec 17 08:41:05 crc kubenswrapper[4966]: I1217 08:41:05.575084 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.171:5353: connect: connection refused" Dec 17 08:41:05 crc kubenswrapper[4966]: I1217 08:41:05.829607 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:41:06 crc kubenswrapper[4966]: I1217 08:41:06.035207 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:41:06 crc kubenswrapper[4966]: I1217 08:41:06.126400 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6cb857dc54-5f9tj" Dec 17 08:41:06 crc kubenswrapper[4966]: I1217 08:41:06.200484 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-68976d496b-zns8c"] Dec 17 08:41:06 crc kubenswrapper[4966]: I1217 08:41:06.320229 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-6c6f84469f-k22bc" Dec 17 08:41:06 crc kubenswrapper[4966]: I1217 08:41:06.412062 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7668494dc4-2sqhx"] Dec 17 08:41:06 crc kubenswrapper[4966]: E1217 08:41:06.860808 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 17 08:41:06 crc kubenswrapper[4966]: E1217 08:41:06.919050 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 17 08:41:06 crc kubenswrapper[4966]: E1217 08:41:06.921135 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 17 08:41:06 crc kubenswrapper[4966]: E1217 08:41:06.921172 4966 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-b67cfd874-bklxp" podUID="f33826e0-0269-4b99-9dce-fcaf62f67bed" containerName="heat-engine" Dec 17 08:41:10 crc kubenswrapper[4966]: I1217 08:41:10.575194 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.171:5353: connect: connection refused" Dec 17 08:41:11 crc kubenswrapper[4966]: I1217 08:41:11.071177 4966 generic.go:334] "Generic (PLEG): container finished" podID="f33826e0-0269-4b99-9dce-fcaf62f67bed" containerID="e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6" exitCode=0 Dec 17 08:41:11 crc kubenswrapper[4966]: I1217 08:41:11.071387 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-b67cfd874-bklxp" event={"ID":"f33826e0-0269-4b99-9dce-fcaf62f67bed","Type":"ContainerDied","Data":"e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6"} Dec 17 08:41:14 crc kubenswrapper[4966]: E1217 08:41:14.496981 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-nova-conductor:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:41:14 crc kubenswrapper[4966]: E1217 08:41:14.497291 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-nova-conductor:c3923531bcda0b0811b2d5053f189beb" Dec 17 08:41:14 crc kubenswrapper[4966]: E1217 08:41:14.497466 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-nova-conductor:c3923531bcda0b0811b2d5053f189beb,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n9j4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-hzwzg_openstack(02f6a226-d49b-4726-8116-4b9ab2837338): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 08:41:14 crc kubenswrapper[4966]: E1217 08:41:14.498640 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" podUID="02f6a226-d49b-4726-8116-4b9ab2837338" Dec 17 08:41:14 crc kubenswrapper[4966]: I1217 08:41:14.695262 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:41:14 crc kubenswrapper[4966]: I1217 08:41:14.816813 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data\") pod \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " Dec 17 08:41:14 crc kubenswrapper[4966]: I1217 08:41:14.817223 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmpsh\" (UniqueName: \"kubernetes.io/projected/cdfc5775-0ec0-433c-9774-0e6a56273b8e-kube-api-access-vmpsh\") pod \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " Dec 17 08:41:14 crc kubenswrapper[4966]: I1217 08:41:14.817732 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data-custom\") pod \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " Dec 17 08:41:14 crc kubenswrapper[4966]: I1217 08:41:14.817935 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-combined-ca-bundle\") pod \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\" (UID: \"cdfc5775-0ec0-433c-9774-0e6a56273b8e\") " Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.530313 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cdfc5775-0ec0-433c-9774-0e6a56273b8e" (UID: "cdfc5775-0ec0-433c-9774-0e6a56273b8e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.533144 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdfc5775-0ec0-433c-9774-0e6a56273b8e-kube-api-access-vmpsh" (OuterVolumeSpecName: "kube-api-access-vmpsh") pod "cdfc5775-0ec0-433c-9774-0e6a56273b8e" (UID: "cdfc5775-0ec0-433c-9774-0e6a56273b8e"). InnerVolumeSpecName "kube-api-access-vmpsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.582331 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.171:5353: connect: connection refused" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.584160 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.600477 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdfc5775-0ec0-433c-9774-0e6a56273b8e" (UID: "cdfc5775-0ec0-433c-9774-0e6a56273b8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.620533 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data-custom\") pod \"f1e350c4-e807-4015-8398-42bad206482f\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.620636 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg2rn\" (UniqueName: \"kubernetes.io/projected/f1e350c4-e807-4015-8398-42bad206482f-kube-api-access-fg2rn\") pod \"f1e350c4-e807-4015-8398-42bad206482f\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.620709 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data\") pod \"f1e350c4-e807-4015-8398-42bad206482f\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.620850 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-combined-ca-bundle\") pod \"f1e350c4-e807-4015-8398-42bad206482f\" (UID: \"f1e350c4-e807-4015-8398-42bad206482f\") " Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.621511 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmpsh\" (UniqueName: \"kubernetes.io/projected/cdfc5775-0ec0-433c-9774-0e6a56273b8e-kube-api-access-vmpsh\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.621538 4966 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.621549 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.640341 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f1e350c4-e807-4015-8398-42bad206482f" (UID: "f1e350c4-e807-4015-8398-42bad206482f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.642981 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e350c4-e807-4015-8398-42bad206482f-kube-api-access-fg2rn" (OuterVolumeSpecName: "kube-api-access-fg2rn") pod "f1e350c4-e807-4015-8398-42bad206482f" (UID: "f1e350c4-e807-4015-8398-42bad206482f"). InnerVolumeSpecName "kube-api-access-fg2rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.646908 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.654294 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-68976d496b-zns8c" Dec 17 08:41:15 crc kubenswrapper[4966]: E1217 08:41:15.658077 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.156:5001/podified-antelope-centos9/openstack-nova-conductor:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" podUID="02f6a226-d49b-4726-8116-4b9ab2837338" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.708129 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data" (OuterVolumeSpecName: "config-data") pod "cdfc5775-0ec0-433c-9774-0e6a56273b8e" (UID: "cdfc5775-0ec0-433c-9774-0e6a56273b8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.737097 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfc5775-0ec0-433c-9774-0e6a56273b8e-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.737126 4966 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.737139 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg2rn\" (UniqueName: \"kubernetes.io/projected/f1e350c4-e807-4015-8398-42bad206482f-kube-api-access-fg2rn\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.770064 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1e350c4-e807-4015-8398-42bad206482f" (UID: "f1e350c4-e807-4015-8398-42bad206482f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.809991 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7668494dc4-2sqhx" event={"ID":"cdfc5775-0ec0-433c-9774-0e6a56273b8e","Type":"ContainerDied","Data":"0c0bd0204551f09f6fdbb12ede1f58faf732ce8c0db0882c3e327efcbfdfe3d1"} Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.810177 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.810248 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-68976d496b-zns8c" event={"ID":"f1e350c4-e807-4015-8398-42bad206482f","Type":"ContainerDied","Data":"3289011c10de198012267280261cb3953329945f3cd0d5cd2f7a06412dcd6f17"} Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.810686 4966 scope.go:117] "RemoveContainer" containerID="001b9b2ac8e52ce9138da62c7bd2d25d5c388b4978e9c26ab219cd343691f642" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.839776 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.848006 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data" (OuterVolumeSpecName: "config-data") pod "f1e350c4-e807-4015-8398-42bad206482f" (UID: "f1e350c4-e807-4015-8398-42bad206482f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.897923 4966 scope.go:117] "RemoveContainer" containerID="aa972212fd850218498cff0b14105afa9d738631c7a214f0dee83d10477ae72e" Dec 17 08:41:15 crc kubenswrapper[4966]: I1217 08:41:15.944592 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e350c4-e807-4015-8398-42bad206482f-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.054276 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.068260 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7668494dc4-2sqhx"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.095617 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7668494dc4-2sqhx"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.119458 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-68976d496b-zns8c"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.136784 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-68976d496b-zns8c"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.200411 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.255129 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.257605 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-swift-storage-0\") pod \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.257733 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-nb\") pod \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.257796 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8d9z\" (UniqueName: \"kubernetes.io/projected/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-kube-api-access-l8d9z\") pod \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.257824 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-config\") pod \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.257953 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-svc\") pod \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.258040 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-sb\") pod \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\" (UID: \"e3e02105-d411-4d80-9d37-ae44d4cd8a4c\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.264718 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-kube-api-access-l8d9z" (OuterVolumeSpecName: "kube-api-access-l8d9z") pod "e3e02105-d411-4d80-9d37-ae44d4cd8a4c" (UID: "e3e02105-d411-4d80-9d37-ae44d4cd8a4c"). InnerVolumeSpecName "kube-api-access-l8d9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.323727 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-config" (OuterVolumeSpecName: "config") pod "e3e02105-d411-4d80-9d37-ae44d4cd8a4c" (UID: "e3e02105-d411-4d80-9d37-ae44d4cd8a4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.341124 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e3e02105-d411-4d80-9d37-ae44d4cd8a4c" (UID: "e3e02105-d411-4d80-9d37-ae44d4cd8a4c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.351985 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e3e02105-d411-4d80-9d37-ae44d4cd8a4c" (UID: "e3e02105-d411-4d80-9d37-ae44d4cd8a4c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.357455 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e3e02105-d411-4d80-9d37-ae44d4cd8a4c" (UID: "e3e02105-d411-4d80-9d37-ae44d4cd8a4c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359509 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data-custom\") pod \"f33826e0-0269-4b99-9dce-fcaf62f67bed\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359539 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-config-data\") pod \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359576 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mfcd\" (UniqueName: \"kubernetes.io/projected/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-kube-api-access-5mfcd\") pod \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359597 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data\") pod \"f33826e0-0269-4b99-9dce-fcaf62f67bed\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359697 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-scripts\") pod \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359732 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-sg-core-conf-yaml\") pod \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359769 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-run-httpd\") pod \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359805 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-combined-ca-bundle\") pod \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359828 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-log-httpd\") pod \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\" (UID: \"67cdcfd5-3af2-475d-bb79-f91357e1b0bd\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359900 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7c6p\" (UniqueName: \"kubernetes.io/projected/f33826e0-0269-4b99-9dce-fcaf62f67bed-kube-api-access-k7c6p\") pod \"f33826e0-0269-4b99-9dce-fcaf62f67bed\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.359964 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-combined-ca-bundle\") pod \"f33826e0-0269-4b99-9dce-fcaf62f67bed\" (UID: \"f33826e0-0269-4b99-9dce-fcaf62f67bed\") " Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.360507 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.360525 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8d9z\" (UniqueName: \"kubernetes.io/projected/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-kube-api-access-l8d9z\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.360535 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.360544 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.360552 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.366794 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-scripts" (OuterVolumeSpecName: "scripts") pod "67cdcfd5-3af2-475d-bb79-f91357e1b0bd" (UID: "67cdcfd5-3af2-475d-bb79-f91357e1b0bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.367823 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "67cdcfd5-3af2-475d-bb79-f91357e1b0bd" (UID: "67cdcfd5-3af2-475d-bb79-f91357e1b0bd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.374287 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "67cdcfd5-3af2-475d-bb79-f91357e1b0bd" (UID: "67cdcfd5-3af2-475d-bb79-f91357e1b0bd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.379306 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f33826e0-0269-4b99-9dce-fcaf62f67bed-kube-api-access-k7c6p" (OuterVolumeSpecName: "kube-api-access-k7c6p") pod "f33826e0-0269-4b99-9dce-fcaf62f67bed" (UID: "f33826e0-0269-4b99-9dce-fcaf62f67bed"). InnerVolumeSpecName "kube-api-access-k7c6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.387684 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-kube-api-access-5mfcd" (OuterVolumeSpecName: "kube-api-access-5mfcd") pod "67cdcfd5-3af2-475d-bb79-f91357e1b0bd" (UID: "67cdcfd5-3af2-475d-bb79-f91357e1b0bd"). InnerVolumeSpecName "kube-api-access-5mfcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.399477 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f33826e0-0269-4b99-9dce-fcaf62f67bed" (UID: "f33826e0-0269-4b99-9dce-fcaf62f67bed"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.402417 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e3e02105-d411-4d80-9d37-ae44d4cd8a4c" (UID: "e3e02105-d411-4d80-9d37-ae44d4cd8a4c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.450007 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data" (OuterVolumeSpecName: "config-data") pod "f33826e0-0269-4b99-9dce-fcaf62f67bed" (UID: "f33826e0-0269-4b99-9dce-fcaf62f67bed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.460052 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "67cdcfd5-3af2-475d-bb79-f91357e1b0bd" (UID: "67cdcfd5-3af2-475d-bb79-f91357e1b0bd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.462716 4966 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.462743 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mfcd\" (UniqueName: \"kubernetes.io/projected/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-kube-api-access-5mfcd\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.462755 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.462763 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.462792 4966 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.462801 4966 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.462809 4966 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.462818 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7c6p\" (UniqueName: \"kubernetes.io/projected/f33826e0-0269-4b99-9dce-fcaf62f67bed-kube-api-access-k7c6p\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.462827 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3e02105-d411-4d80-9d37-ae44d4cd8a4c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.476993 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f33826e0-0269-4b99-9dce-fcaf62f67bed" (UID: "f33826e0-0269-4b99-9dce-fcaf62f67bed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.506678 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67cdcfd5-3af2-475d-bb79-f91357e1b0bd" (UID: "67cdcfd5-3af2-475d-bb79-f91357e1b0bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.532959 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-config-data" (OuterVolumeSpecName: "config-data") pod "67cdcfd5-3af2-475d-bb79-f91357e1b0bd" (UID: "67cdcfd5-3af2-475d-bb79-f91357e1b0bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.565102 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33826e0-0269-4b99-9dce-fcaf62f67bed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.565299 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.565391 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cdcfd5-3af2-475d-bb79-f91357e1b0bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.665525 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-b67cfd874-bklxp" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.665600 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-b67cfd874-bklxp" event={"ID":"f33826e0-0269-4b99-9dce-fcaf62f67bed","Type":"ContainerDied","Data":"d43ee98b23617cef894d5dacee3a49ec79944953e7bdcdbfe07c8d8609c2f57e"} Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.666310 4966 scope.go:117] "RemoveContainer" containerID="e65cc32c190cd51627eeb13d794baafa135bd229de95ec1ec356448d735d28b6" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.688226 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" event={"ID":"e3e02105-d411-4d80-9d37-ae44d4cd8a4c","Type":"ContainerDied","Data":"0851c0256cf554c3732ecb7d2aa6d032f094cc221bb092bc3aec3214caf1e279"} Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.688636 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798df684f5-9xzvn" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.696408 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67cdcfd5-3af2-475d-bb79-f91357e1b0bd","Type":"ContainerDied","Data":"97f1ac08aa28a49a45d392c10460025c7551080a4bbafb8d130bd21a3ce31d33"} Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.696545 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.741544 4966 scope.go:117] "RemoveContainer" containerID="37887270f964c260dd0744dae832d55f04a977e3e67d25a5dbe0c9863d3935dc" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.757222 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-798df684f5-9xzvn"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.780767 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-798df684f5-9xzvn"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.782550 4966 scope.go:117] "RemoveContainer" containerID="6697217dbf113e688e6fd26cbe8283c10b8a4cf3beb83a634921c6f8a5a9916c" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.807331 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.807537 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.807888 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-b67cfd874-bklxp"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.830213 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-b67cfd874-bklxp"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.846112 4966 scope.go:117] "RemoveContainer" containerID="fb31d48108cb02ca483a69aae35f6ba634c4e2999930e746821d372d6717a63d" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.865660 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" path="/var/lib/kubelet/pods/cdfc5775-0ec0-433c-9774-0e6a56273b8e/volumes" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.869767 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" path="/var/lib/kubelet/pods/e3e02105-d411-4d80-9d37-ae44d4cd8a4c/volumes" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.870453 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e350c4-e807-4015-8398-42bad206482f" path="/var/lib/kubelet/pods/f1e350c4-e807-4015-8398-42bad206482f/volumes" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.871635 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f33826e0-0269-4b99-9dce-fcaf62f67bed" path="/var/lib/kubelet/pods/f33826e0-0269-4b99-9dce-fcaf62f67bed/volumes" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.873560 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.876583 4966 scope.go:117] "RemoveContainer" containerID="90efbdbe12e50f6e45148a17f90d31339da3da4aeddd5b010295c46f04149b98" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.897512 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908197 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908574 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e350c4-e807-4015-8398-42bad206482f" containerName="heat-api" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908590 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e350c4-e807-4015-8398-42bad206482f" containerName="heat-api" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908606 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="ceilometer-notification-agent" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908612 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="ceilometer-notification-agent" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908623 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="proxy-httpd" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908629 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="proxy-httpd" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908642 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="ceilometer-central-agent" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908648 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="ceilometer-central-agent" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908662 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e350c4-e807-4015-8398-42bad206482f" containerName="heat-api" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908668 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e350c4-e807-4015-8398-42bad206482f" containerName="heat-api" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908680 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33826e0-0269-4b99-9dce-fcaf62f67bed" containerName="heat-engine" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908685 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33826e0-0269-4b99-9dce-fcaf62f67bed" containerName="heat-engine" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908696 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerName="init" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908701 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerName="init" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908721 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" containerName="heat-cfnapi" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908726 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" containerName="heat-cfnapi" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908738 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="sg-core" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908744 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="sg-core" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908753 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" containerName="heat-cfnapi" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908758 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" containerName="heat-cfnapi" Dec 17 08:41:16 crc kubenswrapper[4966]: E1217 08:41:16.908771 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerName="dnsmasq-dns" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908777 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerName="dnsmasq-dns" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908966 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="proxy-httpd" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908978 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" containerName="heat-cfnapi" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908984 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="sg-core" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.908994 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="ceilometer-notification-agent" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.909004 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e350c4-e807-4015-8398-42bad206482f" containerName="heat-api" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.909013 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e350c4-e807-4015-8398-42bad206482f" containerName="heat-api" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.909021 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f33826e0-0269-4b99-9dce-fcaf62f67bed" containerName="heat-engine" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.909034 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdfc5775-0ec0-433c-9774-0e6a56273b8e" containerName="heat-cfnapi" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.909041 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" containerName="ceilometer-central-agent" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.909052 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e02105-d411-4d80-9d37-ae44d4cd8a4c" containerName="dnsmasq-dns" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.910621 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.913637 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.914419 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.919665 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.928164 4966 scope.go:117] "RemoveContainer" containerID="4f790cb8cb7a6f66eef74cdfe3c1adb0471632552a25bc333210d5ec2718315f" Dec 17 08:41:16 crc kubenswrapper[4966]: I1217 08:41:16.971314 4966 scope.go:117] "RemoveContainer" containerID="35d10ea52b1c854bf8b185fb2c6a83221681162903ab3b1f481412088424d8da" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.072169 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-run-httpd\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.072405 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-config-data\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.072655 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62gws\" (UniqueName: \"kubernetes.io/projected/a6894341-8765-4100-a0d9-c2f8b40f9607-kube-api-access-62gws\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.072815 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-scripts\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.072854 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-log-httpd\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.072944 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.073058 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.174946 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-scripts\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.175001 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-log-httpd\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.175030 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.175069 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.175126 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-run-httpd\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.175154 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-config-data\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.175190 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62gws\" (UniqueName: \"kubernetes.io/projected/a6894341-8765-4100-a0d9-c2f8b40f9607-kube-api-access-62gws\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.176628 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-run-httpd\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.176849 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-log-httpd\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.181184 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-config-data\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.194723 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.195619 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-scripts\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.196241 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.211174 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62gws\" (UniqueName: \"kubernetes.io/projected/a6894341-8765-4100-a0d9-c2f8b40f9607-kube-api-access-62gws\") pod \"ceilometer-0\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.230759 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:17 crc kubenswrapper[4966]: I1217 08:41:17.763551 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:18 crc kubenswrapper[4966]: I1217 08:41:18.775181 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerStarted","Data":"8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89"} Dec 17 08:41:18 crc kubenswrapper[4966]: I1217 08:41:18.775443 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerStarted","Data":"ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983"} Dec 17 08:41:18 crc kubenswrapper[4966]: I1217 08:41:18.775453 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerStarted","Data":"a524f479782fd6499342c20baa06f3f96b12a69ab2b235e9016cd3619a8be51f"} Dec 17 08:41:18 crc kubenswrapper[4966]: I1217 08:41:18.874375 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67cdcfd5-3af2-475d-bb79-f91357e1b0bd" path="/var/lib/kubelet/pods/67cdcfd5-3af2-475d-bb79-f91357e1b0bd/volumes" Dec 17 08:41:19 crc kubenswrapper[4966]: I1217 08:41:19.800945 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerStarted","Data":"f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f"} Dec 17 08:41:21 crc kubenswrapper[4966]: I1217 08:41:21.817189 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerStarted","Data":"854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2"} Dec 17 08:41:21 crc kubenswrapper[4966]: I1217 08:41:21.817933 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 17 08:41:21 crc kubenswrapper[4966]: I1217 08:41:21.847993 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.9946478279999997 podStartE2EDuration="5.847975377s" podCreationTimestamp="2025-12-17 08:41:16 +0000 UTC" firstStartedPulling="2025-12-17 08:41:17.757863879 +0000 UTC m=+1213.302933821" lastFinishedPulling="2025-12-17 08:41:20.611191438 +0000 UTC m=+1216.156261370" observedRunningTime="2025-12-17 08:41:21.839963468 +0000 UTC m=+1217.385033420" watchObservedRunningTime="2025-12-17 08:41:21.847975377 +0000 UTC m=+1217.393045319" Dec 17 08:41:21 crc kubenswrapper[4966]: I1217 08:41:21.969043 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:23 crc kubenswrapper[4966]: I1217 08:41:23.833992 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="ceilometer-central-agent" containerID="cri-o://ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983" gracePeriod=30 Dec 17 08:41:23 crc kubenswrapper[4966]: I1217 08:41:23.834015 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="proxy-httpd" containerID="cri-o://854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2" gracePeriod=30 Dec 17 08:41:23 crc kubenswrapper[4966]: I1217 08:41:23.834076 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="sg-core" containerID="cri-o://f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f" gracePeriod=30 Dec 17 08:41:23 crc kubenswrapper[4966]: I1217 08:41:23.834087 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="ceilometer-notification-agent" containerID="cri-o://8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89" gracePeriod=30 Dec 17 08:41:24 crc kubenswrapper[4966]: I1217 08:41:24.843757 4966 generic.go:334] "Generic (PLEG): container finished" podID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerID="854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2" exitCode=0 Dec 17 08:41:24 crc kubenswrapper[4966]: I1217 08:41:24.844075 4966 generic.go:334] "Generic (PLEG): container finished" podID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerID="f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f" exitCode=2 Dec 17 08:41:24 crc kubenswrapper[4966]: I1217 08:41:24.844085 4966 generic.go:334] "Generic (PLEG): container finished" podID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerID="8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89" exitCode=0 Dec 17 08:41:24 crc kubenswrapper[4966]: I1217 08:41:24.843826 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerDied","Data":"854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2"} Dec 17 08:41:24 crc kubenswrapper[4966]: I1217 08:41:24.844123 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerDied","Data":"f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f"} Dec 17 08:41:24 crc kubenswrapper[4966]: I1217 08:41:24.844136 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerDied","Data":"8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89"} Dec 17 08:41:29 crc kubenswrapper[4966]: I1217 08:41:29.890799 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" event={"ID":"02f6a226-d49b-4726-8116-4b9ab2837338","Type":"ContainerStarted","Data":"21cb6a52d5dbe8f6273ea2d4a3330d298f871767fe3b4b5b28e6a463c09fef72"} Dec 17 08:41:29 crc kubenswrapper[4966]: I1217 08:41:29.921089 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" podStartSLOduration=2.554683514 podStartE2EDuration="32.921063908s" podCreationTimestamp="2025-12-17 08:40:57 +0000 UTC" firstStartedPulling="2025-12-17 08:40:58.521816784 +0000 UTC m=+1194.066886716" lastFinishedPulling="2025-12-17 08:41:28.888197168 +0000 UTC m=+1224.433267110" observedRunningTime="2025-12-17 08:41:29.911175288 +0000 UTC m=+1225.456245230" watchObservedRunningTime="2025-12-17 08:41:29.921063908 +0000 UTC m=+1225.466133870" Dec 17 08:41:31 crc kubenswrapper[4966]: I1217 08:41:31.902363 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:31 crc kubenswrapper[4966]: I1217 08:41:31.912592 4966 generic.go:334] "Generic (PLEG): container finished" podID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerID="ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983" exitCode=0 Dec 17 08:41:31 crc kubenswrapper[4966]: I1217 08:41:31.912639 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerDied","Data":"ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983"} Dec 17 08:41:31 crc kubenswrapper[4966]: I1217 08:41:31.912668 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6894341-8765-4100-a0d9-c2f8b40f9607","Type":"ContainerDied","Data":"a524f479782fd6499342c20baa06f3f96b12a69ab2b235e9016cd3619a8be51f"} Dec 17 08:41:31 crc kubenswrapper[4966]: I1217 08:41:31.912691 4966 scope.go:117] "RemoveContainer" containerID="854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2" Dec 17 08:41:31 crc kubenswrapper[4966]: I1217 08:41:31.912711 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:31 crc kubenswrapper[4966]: I1217 08:41:31.936864 4966 scope.go:117] "RemoveContainer" containerID="f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f" Dec 17 08:41:31 crc kubenswrapper[4966]: I1217 08:41:31.965706 4966 scope.go:117] "RemoveContainer" containerID="8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89" Dec 17 08:41:31 crc kubenswrapper[4966]: I1217 08:41:31.996148 4966 scope.go:117] "RemoveContainer" containerID="ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.012974 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-config-data\") pod \"a6894341-8765-4100-a0d9-c2f8b40f9607\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.013029 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-run-httpd\") pod \"a6894341-8765-4100-a0d9-c2f8b40f9607\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.013049 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-scripts\") pod \"a6894341-8765-4100-a0d9-c2f8b40f9607\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.013154 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-log-httpd\") pod \"a6894341-8765-4100-a0d9-c2f8b40f9607\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.013216 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62gws\" (UniqueName: \"kubernetes.io/projected/a6894341-8765-4100-a0d9-c2f8b40f9607-kube-api-access-62gws\") pod \"a6894341-8765-4100-a0d9-c2f8b40f9607\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.013239 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-combined-ca-bundle\") pod \"a6894341-8765-4100-a0d9-c2f8b40f9607\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.013262 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-sg-core-conf-yaml\") pod \"a6894341-8765-4100-a0d9-c2f8b40f9607\" (UID: \"a6894341-8765-4100-a0d9-c2f8b40f9607\") " Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.020161 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a6894341-8765-4100-a0d9-c2f8b40f9607" (UID: "a6894341-8765-4100-a0d9-c2f8b40f9607"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.020200 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a6894341-8765-4100-a0d9-c2f8b40f9607" (UID: "a6894341-8765-4100-a0d9-c2f8b40f9607"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.059340 4966 scope.go:117] "RemoveContainer" containerID="854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2" Dec 17 08:41:32 crc kubenswrapper[4966]: E1217 08:41:32.059938 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2\": container with ID starting with 854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2 not found: ID does not exist" containerID="854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.059985 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2"} err="failed to get container status \"854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2\": rpc error: code = NotFound desc = could not find container \"854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2\": container with ID starting with 854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2 not found: ID does not exist" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.060012 4966 scope.go:117] "RemoveContainer" containerID="f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.065905 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6894341-8765-4100-a0d9-c2f8b40f9607-kube-api-access-62gws" (OuterVolumeSpecName: "kube-api-access-62gws") pod "a6894341-8765-4100-a0d9-c2f8b40f9607" (UID: "a6894341-8765-4100-a0d9-c2f8b40f9607"). InnerVolumeSpecName "kube-api-access-62gws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.065999 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-scripts" (OuterVolumeSpecName: "scripts") pod "a6894341-8765-4100-a0d9-c2f8b40f9607" (UID: "a6894341-8765-4100-a0d9-c2f8b40f9607"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:32 crc kubenswrapper[4966]: E1217 08:41:32.068729 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f\": container with ID starting with f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f not found: ID does not exist" containerID="f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.068805 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f"} err="failed to get container status \"f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f\": rpc error: code = NotFound desc = could not find container \"f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f\": container with ID starting with f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f not found: ID does not exist" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.068903 4966 scope.go:117] "RemoveContainer" containerID="8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89" Dec 17 08:41:32 crc kubenswrapper[4966]: E1217 08:41:32.069532 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89\": container with ID starting with 8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89 not found: ID does not exist" containerID="8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.069606 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89"} err="failed to get container status \"8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89\": rpc error: code = NotFound desc = could not find container \"8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89\": container with ID starting with 8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89 not found: ID does not exist" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.069662 4966 scope.go:117] "RemoveContainer" containerID="ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983" Dec 17 08:41:32 crc kubenswrapper[4966]: E1217 08:41:32.070431 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983\": container with ID starting with ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983 not found: ID does not exist" containerID="ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.070489 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983"} err="failed to get container status \"ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983\": rpc error: code = NotFound desc = could not find container \"ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983\": container with ID starting with ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983 not found: ID does not exist" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.101104 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a6894341-8765-4100-a0d9-c2f8b40f9607" (UID: "a6894341-8765-4100-a0d9-c2f8b40f9607"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.116046 4966 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.116099 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62gws\" (UniqueName: \"kubernetes.io/projected/a6894341-8765-4100-a0d9-c2f8b40f9607-kube-api-access-62gws\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.116112 4966 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.116120 4966 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6894341-8765-4100-a0d9-c2f8b40f9607-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.116129 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.164802 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6894341-8765-4100-a0d9-c2f8b40f9607" (UID: "a6894341-8765-4100-a0d9-c2f8b40f9607"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.202764 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-config-data" (OuterVolumeSpecName: "config-data") pod "a6894341-8765-4100-a0d9-c2f8b40f9607" (UID: "a6894341-8765-4100-a0d9-c2f8b40f9607"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.219007 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.219041 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6894341-8765-4100-a0d9-c2f8b40f9607-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.253888 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.263456 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.287237 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:32 crc kubenswrapper[4966]: E1217 08:41:32.287743 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="proxy-httpd" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.287767 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="proxy-httpd" Dec 17 08:41:32 crc kubenswrapper[4966]: E1217 08:41:32.287802 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="sg-core" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.287810 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="sg-core" Dec 17 08:41:32 crc kubenswrapper[4966]: E1217 08:41:32.287823 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="ceilometer-notification-agent" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.287831 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="ceilometer-notification-agent" Dec 17 08:41:32 crc kubenswrapper[4966]: E1217 08:41:32.287862 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="ceilometer-central-agent" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.287886 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="ceilometer-central-agent" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.288137 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="ceilometer-central-agent" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.288155 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="ceilometer-notification-agent" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.288173 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="sg-core" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.288201 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" containerName="proxy-httpd" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.290414 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.293226 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.293569 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.300394 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.425395 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.425433 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-log-httpd\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.425559 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwfhz\" (UniqueName: \"kubernetes.io/projected/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-kube-api-access-xwfhz\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.425586 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-config-data\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.425610 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-scripts\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.425677 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-run-httpd\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.425732 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.528546 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.528601 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-log-httpd\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.528805 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwfhz\" (UniqueName: \"kubernetes.io/projected/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-kube-api-access-xwfhz\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.528839 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-config-data\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.528902 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-scripts\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.528956 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-run-httpd\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.529001 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.529171 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-log-httpd\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.529903 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-run-httpd\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.533130 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.533693 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-config-data\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.534704 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.535674 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-scripts\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.547159 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwfhz\" (UniqueName: \"kubernetes.io/projected/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-kube-api-access-xwfhz\") pod \"ceilometer-0\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.606213 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:32 crc kubenswrapper[4966]: I1217 08:41:32.854294 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6894341-8765-4100-a0d9-c2f8b40f9607" path="/var/lib/kubelet/pods/a6894341-8765-4100-a0d9-c2f8b40f9607/volumes" Dec 17 08:41:33 crc kubenswrapper[4966]: I1217 08:41:33.246296 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:33 crc kubenswrapper[4966]: I1217 08:41:33.946641 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerStarted","Data":"7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a"} Dec 17 08:41:33 crc kubenswrapper[4966]: I1217 08:41:33.946978 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerStarted","Data":"4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6"} Dec 17 08:41:33 crc kubenswrapper[4966]: I1217 08:41:33.947028 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerStarted","Data":"93b8763c03cbc042654ff7ca61e47ba9235bb07963c741e21c2b0515e48980ca"} Dec 17 08:41:34 crc kubenswrapper[4966]: I1217 08:41:34.958609 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerStarted","Data":"966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0"} Dec 17 08:41:35 crc kubenswrapper[4966]: I1217 08:41:35.375506 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:36 crc kubenswrapper[4966]: I1217 08:41:36.995105 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerStarted","Data":"2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f"} Dec 17 08:41:36 crc kubenswrapper[4966]: I1217 08:41:36.995532 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="ceilometer-central-agent" containerID="cri-o://4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6" gracePeriod=30 Dec 17 08:41:36 crc kubenswrapper[4966]: I1217 08:41:36.995783 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 17 08:41:36 crc kubenswrapper[4966]: I1217 08:41:36.996046 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="proxy-httpd" containerID="cri-o://2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f" gracePeriod=30 Dec 17 08:41:36 crc kubenswrapper[4966]: I1217 08:41:36.996092 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="sg-core" containerID="cri-o://966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0" gracePeriod=30 Dec 17 08:41:36 crc kubenswrapper[4966]: I1217 08:41:36.996125 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="ceilometer-notification-agent" containerID="cri-o://7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a" gracePeriod=30 Dec 17 08:41:37 crc kubenswrapper[4966]: I1217 08:41:37.027818 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.848395965 podStartE2EDuration="5.027803332s" podCreationTimestamp="2025-12-17 08:41:32 +0000 UTC" firstStartedPulling="2025-12-17 08:41:33.248324987 +0000 UTC m=+1228.793394929" lastFinishedPulling="2025-12-17 08:41:36.427732354 +0000 UTC m=+1231.972802296" observedRunningTime="2025-12-17 08:41:37.022550138 +0000 UTC m=+1232.567620080" watchObservedRunningTime="2025-12-17 08:41:37.027803332 +0000 UTC m=+1232.572873274" Dec 17 08:41:38 crc kubenswrapper[4966]: I1217 08:41:38.006356 4966 generic.go:334] "Generic (PLEG): container finished" podID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerID="2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f" exitCode=0 Dec 17 08:41:38 crc kubenswrapper[4966]: I1217 08:41:38.006687 4966 generic.go:334] "Generic (PLEG): container finished" podID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerID="966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0" exitCode=2 Dec 17 08:41:38 crc kubenswrapper[4966]: I1217 08:41:38.006702 4966 generic.go:334] "Generic (PLEG): container finished" podID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerID="7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a" exitCode=0 Dec 17 08:41:38 crc kubenswrapper[4966]: I1217 08:41:38.006409 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerDied","Data":"2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f"} Dec 17 08:41:38 crc kubenswrapper[4966]: I1217 08:41:38.006755 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerDied","Data":"966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0"} Dec 17 08:41:38 crc kubenswrapper[4966]: I1217 08:41:38.006768 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerDied","Data":"7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a"} Dec 17 08:41:38 crc kubenswrapper[4966]: I1217 08:41:38.390093 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:41:38 crc kubenswrapper[4966]: I1217 08:41:38.390378 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerName="glance-log" containerID="cri-o://b0d8bf408d4225ebf88e7dd96f46b53365c4bc3cacf91efa323ad0b6e873dabd" gracePeriod=30 Dec 17 08:41:38 crc kubenswrapper[4966]: I1217 08:41:38.390542 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerName="glance-httpd" containerID="cri-o://0dee864237b1ca3a178e7b9abb0c4951cf10120ef5333b8af086c605dd1a4ad6" gracePeriod=30 Dec 17 08:41:39 crc kubenswrapper[4966]: I1217 08:41:39.018422 4966 generic.go:334] "Generic (PLEG): container finished" podID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerID="b0d8bf408d4225ebf88e7dd96f46b53365c4bc3cacf91efa323ad0b6e873dabd" exitCode=143 Dec 17 08:41:39 crc kubenswrapper[4966]: I1217 08:41:39.018499 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7a7b745-08ce-4be8-b81b-82f632be09a3","Type":"ContainerDied","Data":"b0d8bf408d4225ebf88e7dd96f46b53365c4bc3cacf91efa323ad0b6e873dabd"} Dec 17 08:41:39 crc kubenswrapper[4966]: I1217 08:41:39.864760 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:41:39 crc kubenswrapper[4966]: I1217 08:41:39.865414 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="073e8454-1a1f-4a48-9516-070e0fddc046" containerName="glance-log" containerID="cri-o://14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119" gracePeriod=30 Dec 17 08:41:39 crc kubenswrapper[4966]: I1217 08:41:39.865750 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="073e8454-1a1f-4a48-9516-070e0fddc046" containerName="glance-httpd" containerID="cri-o://9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc" gracePeriod=30 Dec 17 08:41:40 crc kubenswrapper[4966]: I1217 08:41:40.034496 4966 generic.go:334] "Generic (PLEG): container finished" podID="073e8454-1a1f-4a48-9516-070e0fddc046" containerID="14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119" exitCode=143 Dec 17 08:41:40 crc kubenswrapper[4966]: I1217 08:41:40.034560 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"073e8454-1a1f-4a48-9516-070e0fddc046","Type":"ContainerDied","Data":"14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119"} Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.053300 4966 generic.go:334] "Generic (PLEG): container finished" podID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerID="0dee864237b1ca3a178e7b9abb0c4951cf10120ef5333b8af086c605dd1a4ad6" exitCode=0 Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.053592 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7a7b745-08ce-4be8-b81b-82f632be09a3","Type":"ContainerDied","Data":"0dee864237b1ca3a178e7b9abb0c4951cf10120ef5333b8af086c605dd1a4ad6"} Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.193817 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.235237 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w427r\" (UniqueName: \"kubernetes.io/projected/d7a7b745-08ce-4be8-b81b-82f632be09a3-kube-api-access-w427r\") pod \"d7a7b745-08ce-4be8-b81b-82f632be09a3\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.235291 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-scripts\") pod \"d7a7b745-08ce-4be8-b81b-82f632be09a3\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.235357 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"d7a7b745-08ce-4be8-b81b-82f632be09a3\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.235390 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-public-tls-certs\") pod \"d7a7b745-08ce-4be8-b81b-82f632be09a3\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.235668 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-logs\") pod \"d7a7b745-08ce-4be8-b81b-82f632be09a3\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.235702 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-httpd-run\") pod \"d7a7b745-08ce-4be8-b81b-82f632be09a3\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.235747 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-config-data\") pod \"d7a7b745-08ce-4be8-b81b-82f632be09a3\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.235774 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-combined-ca-bundle\") pod \"d7a7b745-08ce-4be8-b81b-82f632be09a3\" (UID: \"d7a7b745-08ce-4be8-b81b-82f632be09a3\") " Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.238959 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-logs" (OuterVolumeSpecName: "logs") pod "d7a7b745-08ce-4be8-b81b-82f632be09a3" (UID: "d7a7b745-08ce-4be8-b81b-82f632be09a3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.239344 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d7a7b745-08ce-4be8-b81b-82f632be09a3" (UID: "d7a7b745-08ce-4be8-b81b-82f632be09a3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.244767 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-scripts" (OuterVolumeSpecName: "scripts") pod "d7a7b745-08ce-4be8-b81b-82f632be09a3" (UID: "d7a7b745-08ce-4be8-b81b-82f632be09a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.259702 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7a7b745-08ce-4be8-b81b-82f632be09a3-kube-api-access-w427r" (OuterVolumeSpecName: "kube-api-access-w427r") pod "d7a7b745-08ce-4be8-b81b-82f632be09a3" (UID: "d7a7b745-08ce-4be8-b81b-82f632be09a3"). InnerVolumeSpecName "kube-api-access-w427r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.268320 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "d7a7b745-08ce-4be8-b81b-82f632be09a3" (UID: "d7a7b745-08ce-4be8-b81b-82f632be09a3"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.319335 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d7a7b745-08ce-4be8-b81b-82f632be09a3" (UID: "d7a7b745-08ce-4be8-b81b-82f632be09a3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.348995 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w427r\" (UniqueName: \"kubernetes.io/projected/d7a7b745-08ce-4be8-b81b-82f632be09a3-kube-api-access-w427r\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.349034 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.349092 4966 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.349106 4966 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.349120 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.349132 4966 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7a7b745-08ce-4be8-b81b-82f632be09a3-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.354422 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-config-data" (OuterVolumeSpecName: "config-data") pod "d7a7b745-08ce-4be8-b81b-82f632be09a3" (UID: "d7a7b745-08ce-4be8-b81b-82f632be09a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.381365 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7a7b745-08ce-4be8-b81b-82f632be09a3" (UID: "d7a7b745-08ce-4be8-b81b-82f632be09a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.391909 4966 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.450820 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.450860 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a7b745-08ce-4be8-b81b-82f632be09a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:42 crc kubenswrapper[4966]: I1217 08:41:42.450895 4966 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.062792 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7a7b745-08ce-4be8-b81b-82f632be09a3","Type":"ContainerDied","Data":"8edf28b9737ddca025b20f80661e6bac0063e79b9f49e79027b45c0f9f760998"} Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.063081 4966 scope.go:117] "RemoveContainer" containerID="0dee864237b1ca3a178e7b9abb0c4951cf10120ef5333b8af086c605dd1a4ad6" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.063134 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.169604 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.176973 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.180177 4966 scope.go:117] "RemoveContainer" containerID="b0d8bf408d4225ebf88e7dd96f46b53365c4bc3cacf91efa323ad0b6e873dabd" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.207802 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:41:43 crc kubenswrapper[4966]: E1217 08:41:43.208805 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerName="glance-httpd" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.208821 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerName="glance-httpd" Dec 17 08:41:43 crc kubenswrapper[4966]: E1217 08:41:43.208853 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerName="glance-log" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.208859 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerName="glance-log" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.209062 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerName="glance-log" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.209092 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a7b745-08ce-4be8-b81b-82f632be09a3" containerName="glance-httpd" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.210072 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.217840 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.218429 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.235068 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.371561 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.371616 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99wrs\" (UniqueName: \"kubernetes.io/projected/2e9200b2-073f-42c9-a997-dfe8131670ff-kube-api-access-99wrs\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.371683 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.371734 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-config-data\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.371760 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-scripts\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.371802 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.371822 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e9200b2-073f-42c9-a997-dfe8131670ff-logs\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.371853 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e9200b2-073f-42c9-a997-dfe8131670ff-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.474493 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.474565 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99wrs\" (UniqueName: \"kubernetes.io/projected/2e9200b2-073f-42c9-a997-dfe8131670ff-kube-api-access-99wrs\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.474609 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.474668 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-config-data\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.474713 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-scripts\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.474792 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.474814 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e9200b2-073f-42c9-a997-dfe8131670ff-logs\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.474863 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e9200b2-073f-42c9-a997-dfe8131670ff-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.475435 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e9200b2-073f-42c9-a997-dfe8131670ff-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.477792 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e9200b2-073f-42c9-a997-dfe8131670ff-logs\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.479535 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.484669 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-scripts\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.488101 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-config-data\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.489687 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.498520 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99wrs\" (UniqueName: \"kubernetes.io/projected/2e9200b2-073f-42c9-a997-dfe8131670ff-kube-api-access-99wrs\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.529230 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e9200b2-073f-42c9-a997-dfe8131670ff-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.630433 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2e9200b2-073f-42c9-a997-dfe8131670ff\") " pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.772270 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.833285 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.889713 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-logs\") pod \"073e8454-1a1f-4a48-9516-070e0fddc046\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.889982 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-scripts\") pod \"073e8454-1a1f-4a48-9516-070e0fddc046\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.890033 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-httpd-run\") pod \"073e8454-1a1f-4a48-9516-070e0fddc046\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.890057 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbhl4\" (UniqueName: \"kubernetes.io/projected/073e8454-1a1f-4a48-9516-070e0fddc046-kube-api-access-hbhl4\") pod \"073e8454-1a1f-4a48-9516-070e0fddc046\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.890092 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"073e8454-1a1f-4a48-9516-070e0fddc046\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.890111 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-config-data\") pod \"073e8454-1a1f-4a48-9516-070e0fddc046\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.890125 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-combined-ca-bundle\") pod \"073e8454-1a1f-4a48-9516-070e0fddc046\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.890182 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-internal-tls-certs\") pod \"073e8454-1a1f-4a48-9516-070e0fddc046\" (UID: \"073e8454-1a1f-4a48-9516-070e0fddc046\") " Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.897677 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "073e8454-1a1f-4a48-9516-070e0fddc046" (UID: "073e8454-1a1f-4a48-9516-070e0fddc046"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.897883 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-logs" (OuterVolumeSpecName: "logs") pod "073e8454-1a1f-4a48-9516-070e0fddc046" (UID: "073e8454-1a1f-4a48-9516-070e0fddc046"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.901239 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "073e8454-1a1f-4a48-9516-070e0fddc046" (UID: "073e8454-1a1f-4a48-9516-070e0fddc046"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.901370 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/073e8454-1a1f-4a48-9516-070e0fddc046-kube-api-access-hbhl4" (OuterVolumeSpecName: "kube-api-access-hbhl4") pod "073e8454-1a1f-4a48-9516-070e0fddc046" (UID: "073e8454-1a1f-4a48-9516-070e0fddc046"). InnerVolumeSpecName "kube-api-access-hbhl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.912296 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-scripts" (OuterVolumeSpecName: "scripts") pod "073e8454-1a1f-4a48-9516-070e0fddc046" (UID: "073e8454-1a1f-4a48-9516-070e0fddc046"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.991459 4966 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.991487 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbhl4\" (UniqueName: \"kubernetes.io/projected/073e8454-1a1f-4a48-9516-070e0fddc046-kube-api-access-hbhl4\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.991516 4966 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.991525 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/073e8454-1a1f-4a48-9516-070e0fddc046-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:43 crc kubenswrapper[4966]: I1217 08:41:43.991533 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.021279 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "073e8454-1a1f-4a48-9516-070e0fddc046" (UID: "073e8454-1a1f-4a48-9516-070e0fddc046"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.024141 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "073e8454-1a1f-4a48-9516-070e0fddc046" (UID: "073e8454-1a1f-4a48-9516-070e0fddc046"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.038451 4966 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.061242 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-config-data" (OuterVolumeSpecName: "config-data") pod "073e8454-1a1f-4a48-9516-070e0fddc046" (UID: "073e8454-1a1f-4a48-9516-070e0fddc046"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.093188 4966 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.093214 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.093224 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.093235 4966 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/073e8454-1a1f-4a48-9516-070e0fddc046-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.096115 4966 generic.go:334] "Generic (PLEG): container finished" podID="073e8454-1a1f-4a48-9516-070e0fddc046" containerID="9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc" exitCode=0 Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.096160 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"073e8454-1a1f-4a48-9516-070e0fddc046","Type":"ContainerDied","Data":"9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc"} Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.096184 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"073e8454-1a1f-4a48-9516-070e0fddc046","Type":"ContainerDied","Data":"75ab3382d75a5398870a29868b8a89564dfbb4e02a9cd6e837f5ea34d310a8c0"} Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.096201 4966 scope.go:117] "RemoveContainer" containerID="9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.096318 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.152155 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.156575 4966 scope.go:117] "RemoveContainer" containerID="14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.175135 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.257935 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:41:44 crc kubenswrapper[4966]: E1217 08:41:44.258698 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073e8454-1a1f-4a48-9516-070e0fddc046" containerName="glance-log" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.258710 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="073e8454-1a1f-4a48-9516-070e0fddc046" containerName="glance-log" Dec 17 08:41:44 crc kubenswrapper[4966]: E1217 08:41:44.258738 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073e8454-1a1f-4a48-9516-070e0fddc046" containerName="glance-httpd" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.258744 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="073e8454-1a1f-4a48-9516-070e0fddc046" containerName="glance-httpd" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.258961 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="073e8454-1a1f-4a48-9516-070e0fddc046" containerName="glance-log" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.258987 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="073e8454-1a1f-4a48-9516-070e0fddc046" containerName="glance-httpd" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.259964 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.281562 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.281823 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.286609 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.302054 4966 scope.go:117] "RemoveContainer" containerID="9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc" Dec 17 08:41:44 crc kubenswrapper[4966]: E1217 08:41:44.316021 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc\": container with ID starting with 9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc not found: ID does not exist" containerID="9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.316063 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc"} err="failed to get container status \"9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc\": rpc error: code = NotFound desc = could not find container \"9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc\": container with ID starting with 9186c29b50cf4cb184a97a959c617205accd10ab5fa82dde67ed1316ffe7c9fc not found: ID does not exist" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.316088 4966 scope.go:117] "RemoveContainer" containerID="14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119" Dec 17 08:41:44 crc kubenswrapper[4966]: E1217 08:41:44.321165 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119\": container with ID starting with 14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119 not found: ID does not exist" containerID="14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.321213 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119"} err="failed to get container status \"14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119\": rpc error: code = NotFound desc = could not find container \"14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119\": container with ID starting with 14f5b34975359c2145c5a343dc45000be4e7e4a2982fda07c87270f4fa8cd119 not found: ID does not exist" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.409958 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-logs\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.410005 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.410147 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ktr\" (UniqueName: \"kubernetes.io/projected/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-kube-api-access-j9ktr\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.410209 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.410236 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.410277 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.410351 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.410390 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.512280 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-logs\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.512318 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.512351 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ktr\" (UniqueName: \"kubernetes.io/projected/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-kube-api-access-j9ktr\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.512373 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.512388 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.512409 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.512435 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.512455 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.517328 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.517538 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-logs\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.517855 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.523991 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.527539 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.527603 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.529606 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.536335 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ktr\" (UniqueName: \"kubernetes.io/projected/748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0-kube-api-access-j9ktr\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.554662 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0\") " pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.622463 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.893408 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="073e8454-1a1f-4a48-9516-070e0fddc046" path="/var/lib/kubelet/pods/073e8454-1a1f-4a48-9516-070e0fddc046/volumes" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.897985 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7a7b745-08ce-4be8-b81b-82f632be09a3" path="/var/lib/kubelet/pods/d7a7b745-08ce-4be8-b81b-82f632be09a3/volumes" Dec 17 08:41:44 crc kubenswrapper[4966]: I1217 08:41:44.898911 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 17 08:41:45 crc kubenswrapper[4966]: I1217 08:41:45.119737 4966 generic.go:334] "Generic (PLEG): container finished" podID="02f6a226-d49b-4726-8116-4b9ab2837338" containerID="21cb6a52d5dbe8f6273ea2d4a3330d298f871767fe3b4b5b28e6a463c09fef72" exitCode=0 Dec 17 08:41:45 crc kubenswrapper[4966]: I1217 08:41:45.120064 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" event={"ID":"02f6a226-d49b-4726-8116-4b9ab2837338","Type":"ContainerDied","Data":"21cb6a52d5dbe8f6273ea2d4a3330d298f871767fe3b4b5b28e6a463c09fef72"} Dec 17 08:41:45 crc kubenswrapper[4966]: I1217 08:41:45.125073 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e9200b2-073f-42c9-a997-dfe8131670ff","Type":"ContainerStarted","Data":"70790cafa457565c3425b9e6758a3d38b9159d1abff3224ad7b638ef74a77ef3"} Dec 17 08:41:45 crc kubenswrapper[4966]: I1217 08:41:45.190406 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.170703 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e9200b2-073f-42c9-a997-dfe8131670ff","Type":"ContainerStarted","Data":"0962ef31e1f8eec91b4ecf24b740379ba8a8d6b50018e1849aef52bf4323edfc"} Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.197607 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0","Type":"ContainerStarted","Data":"635eb10c5a7f9e1028b4b8ebd43c254df05e8969832c6d1b434d6bb49ca48621"} Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.601575 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.728402 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.748488 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-config-data\") pod \"02f6a226-d49b-4726-8116-4b9ab2837338\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.748564 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-scripts\") pod \"02f6a226-d49b-4726-8116-4b9ab2837338\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.748624 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9j4c\" (UniqueName: \"kubernetes.io/projected/02f6a226-d49b-4726-8116-4b9ab2837338-kube-api-access-n9j4c\") pod \"02f6a226-d49b-4726-8116-4b9ab2837338\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.748742 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-combined-ca-bundle\") pod \"02f6a226-d49b-4726-8116-4b9ab2837338\" (UID: \"02f6a226-d49b-4726-8116-4b9ab2837338\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.776001 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f6a226-d49b-4726-8116-4b9ab2837338-kube-api-access-n9j4c" (OuterVolumeSpecName: "kube-api-access-n9j4c") pod "02f6a226-d49b-4726-8116-4b9ab2837338" (UID: "02f6a226-d49b-4726-8116-4b9ab2837338"). InnerVolumeSpecName "kube-api-access-n9j4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.782152 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-scripts" (OuterVolumeSpecName: "scripts") pod "02f6a226-d49b-4726-8116-4b9ab2837338" (UID: "02f6a226-d49b-4726-8116-4b9ab2837338"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.808283 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.808330 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.855760 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-config-data" (OuterVolumeSpecName: "config-data") pod "02f6a226-d49b-4726-8116-4b9ab2837338" (UID: "02f6a226-d49b-4726-8116-4b9ab2837338"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.856649 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-config-data\") pod \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.856696 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-run-httpd\") pod \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.856738 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-log-httpd\") pod \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.856761 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-scripts\") pod \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.856831 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-sg-core-conf-yaml\") pod \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.856945 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwfhz\" (UniqueName: \"kubernetes.io/projected/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-kube-api-access-xwfhz\") pod \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.857024 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-combined-ca-bundle\") pod \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\" (UID: \"2e49dad7-7eb2-43bb-afcf-068e4b517ab4\") " Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.858673 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2e49dad7-7eb2-43bb-afcf-068e4b517ab4" (UID: "2e49dad7-7eb2-43bb-afcf-068e4b517ab4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.860809 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02f6a226-d49b-4726-8116-4b9ab2837338" (UID: "02f6a226-d49b-4726-8116-4b9ab2837338"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.861134 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2e49dad7-7eb2-43bb-afcf-068e4b517ab4" (UID: "2e49dad7-7eb2-43bb-afcf-068e4b517ab4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.865120 4966 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.865144 4966 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.865157 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.865169 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.865180 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9j4c\" (UniqueName: \"kubernetes.io/projected/02f6a226-d49b-4726-8116-4b9ab2837338-kube-api-access-n9j4c\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.865190 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6a226-d49b-4726-8116-4b9ab2837338-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.878050 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-kube-api-access-xwfhz" (OuterVolumeSpecName: "kube-api-access-xwfhz") pod "2e49dad7-7eb2-43bb-afcf-068e4b517ab4" (UID: "2e49dad7-7eb2-43bb-afcf-068e4b517ab4"). InnerVolumeSpecName "kube-api-access-xwfhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.915000 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-scripts" (OuterVolumeSpecName: "scripts") pod "2e49dad7-7eb2-43bb-afcf-068e4b517ab4" (UID: "2e49dad7-7eb2-43bb-afcf-068e4b517ab4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.967106 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwfhz\" (UniqueName: \"kubernetes.io/projected/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-kube-api-access-xwfhz\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.967136 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:46 crc kubenswrapper[4966]: I1217 08:41:46.996142 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2e49dad7-7eb2-43bb-afcf-068e4b517ab4" (UID: "2e49dad7-7eb2-43bb-afcf-068e4b517ab4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.050647 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e49dad7-7eb2-43bb-afcf-068e4b517ab4" (UID: "2e49dad7-7eb2-43bb-afcf-068e4b517ab4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.068819 4966 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.068847 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.180121 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-config-data" (OuterVolumeSpecName: "config-data") pod "2e49dad7-7eb2-43bb-afcf-068e4b517ab4" (UID: "2e49dad7-7eb2-43bb-afcf-068e4b517ab4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.207634 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e9200b2-073f-42c9-a997-dfe8131670ff","Type":"ContainerStarted","Data":"42c8f122c757e6c00525bc9d36d3c6c90c64dfdbddcf7d72f9775a4276da97de"} Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.210235 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0","Type":"ContainerStarted","Data":"e316d44858b1b912b52534689d98a049952b0a9427546a4eb8b4882a798de18e"} Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.215624 4966 generic.go:334] "Generic (PLEG): container finished" podID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerID="4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6" exitCode=0 Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.215695 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerDied","Data":"4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6"} Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.215724 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e49dad7-7eb2-43bb-afcf-068e4b517ab4","Type":"ContainerDied","Data":"93b8763c03cbc042654ff7ca61e47ba9235bb07963c741e21c2b0515e48980ca"} Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.215744 4966 scope.go:117] "RemoveContainer" containerID="2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.215940 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.224383 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" event={"ID":"02f6a226-d49b-4726-8116-4b9ab2837338","Type":"ContainerDied","Data":"930ffc4b129f2613ecb947f56b3691c7fb52bcd33d1991639d4acb6994200418"} Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.224421 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="930ffc4b129f2613ecb947f56b3691c7fb52bcd33d1991639d4acb6994200418" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.224735 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-hzwzg" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.247970 4966 scope.go:117] "RemoveContainer" containerID="966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.266044 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.266028686 podStartE2EDuration="4.266028686s" podCreationTimestamp="2025-12-17 08:41:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:41:47.237365539 +0000 UTC m=+1242.782435501" watchObservedRunningTime="2025-12-17 08:41:47.266028686 +0000 UTC m=+1242.811098628" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.272392 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e49dad7-7eb2-43bb-afcf-068e4b517ab4-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.283747 4966 scope.go:117] "RemoveContainer" containerID="7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287011 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.287425 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="proxy-httpd" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287441 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="proxy-httpd" Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.287462 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="ceilometer-central-agent" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287469 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="ceilometer-central-agent" Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.287493 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="ceilometer-notification-agent" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287499 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="ceilometer-notification-agent" Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.287521 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="sg-core" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287527 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="sg-core" Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.287546 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f6a226-d49b-4726-8116-4b9ab2837338" containerName="nova-cell0-conductor-db-sync" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287552 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f6a226-d49b-4726-8116-4b9ab2837338" containerName="nova-cell0-conductor-db-sync" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287736 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="sg-core" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287747 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="proxy-httpd" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287759 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f6a226-d49b-4726-8116-4b9ab2837338" containerName="nova-cell0-conductor-db-sync" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287773 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="ceilometer-notification-agent" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.287783 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" containerName="ceilometer-central-agent" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.288380 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.291349 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.293291 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-s6hbj" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.298217 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.318374 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.334446 4966 scope.go:117] "RemoveContainer" containerID="4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.339008 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.357273 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.360163 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.365903 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.366538 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.375601 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbsms\" (UniqueName: \"kubernetes.io/projected/dc23d16c-1ee3-4ee5-a102-e72e1571c032-kube-api-access-jbsms\") pod \"nova-cell0-conductor-0\" (UID: \"dc23d16c-1ee3-4ee5-a102-e72e1571c032\") " pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.375785 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc23d16c-1ee3-4ee5-a102-e72e1571c032-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dc23d16c-1ee3-4ee5-a102-e72e1571c032\") " pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.375820 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc23d16c-1ee3-4ee5-a102-e72e1571c032-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dc23d16c-1ee3-4ee5-a102-e72e1571c032\") " pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.378003 4966 scope.go:117] "RemoveContainer" containerID="2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f" Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.378415 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f\": container with ID starting with 2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f not found: ID does not exist" containerID="2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.378451 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f"} err="failed to get container status \"2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f\": rpc error: code = NotFound desc = could not find container \"2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f\": container with ID starting with 2df3313ad1f81b71f26b1c8bf77306560460b330e872fbfe2e0c15de86fb9f8f not found: ID does not exist" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.378476 4966 scope.go:117] "RemoveContainer" containerID="966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0" Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.378686 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0\": container with ID starting with 966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0 not found: ID does not exist" containerID="966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.378706 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0"} err="failed to get container status \"966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0\": rpc error: code = NotFound desc = could not find container \"966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0\": container with ID starting with 966566085be05329ae136893708af73faf95c91a63d1604ea4d07a509054aae0 not found: ID does not exist" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.378721 4966 scope.go:117] "RemoveContainer" containerID="7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a" Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.378928 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a\": container with ID starting with 7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a not found: ID does not exist" containerID="7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.378944 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a"} err="failed to get container status \"7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a\": rpc error: code = NotFound desc = could not find container \"7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a\": container with ID starting with 7a1c37f5748b682b779187cd22f7da323e5ced22bc5747efd110ee8425c9eb2a not found: ID does not exist" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.378957 4966 scope.go:117] "RemoveContainer" containerID="4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6" Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.379183 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6\": container with ID starting with 4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6 not found: ID does not exist" containerID="4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.379205 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6"} err="failed to get container status \"4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6\": rpc error: code = NotFound desc = could not find container \"4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6\": container with ID starting with 4b8d4a76c53a623ec445c929f97e888c49d9794e20729d20edde708fe781acb6 not found: ID does not exist" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.382590 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.478119 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc23d16c-1ee3-4ee5-a102-e72e1571c032-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dc23d16c-1ee3-4ee5-a102-e72e1571c032\") " pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.478161 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.478196 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-run-httpd\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.478230 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.478601 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbbk7\" (UniqueName: \"kubernetes.io/projected/c0f67424-421d-4dc9-baad-af7ad452b801-kube-api-access-dbbk7\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.478729 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-log-httpd\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.478788 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbsms\" (UniqueName: \"kubernetes.io/projected/dc23d16c-1ee3-4ee5-a102-e72e1571c032-kube-api-access-jbsms\") pod \"nova-cell0-conductor-0\" (UID: \"dc23d16c-1ee3-4ee5-a102-e72e1571c032\") " pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.478861 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-scripts\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.478977 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc23d16c-1ee3-4ee5-a102-e72e1571c032-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dc23d16c-1ee3-4ee5-a102-e72e1571c032\") " pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.479012 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-config-data\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.483529 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc23d16c-1ee3-4ee5-a102-e72e1571c032-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dc23d16c-1ee3-4ee5-a102-e72e1571c032\") " pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.485499 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc23d16c-1ee3-4ee5-a102-e72e1571c032-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dc23d16c-1ee3-4ee5-a102-e72e1571c032\") " pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.497237 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbsms\" (UniqueName: \"kubernetes.io/projected/dc23d16c-1ee3-4ee5-a102-e72e1571c032-kube-api-access-jbsms\") pod \"nova-cell0-conductor-0\" (UID: \"dc23d16c-1ee3-4ee5-a102-e72e1571c032\") " pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.580397 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-scripts\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.580475 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-config-data\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.580499 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.580529 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-run-httpd\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.580565 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.580652 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbbk7\" (UniqueName: \"kubernetes.io/projected/c0f67424-421d-4dc9-baad-af7ad452b801-kube-api-access-dbbk7\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.580705 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-log-httpd\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.581187 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-log-httpd\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.584332 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-run-httpd\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.588402 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-scripts\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.589420 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-config-data\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.589596 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.595143 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.604637 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbbk7\" (UniqueName: \"kubernetes.io/projected/c0f67424-421d-4dc9-baad-af7ad452b801-kube-api-access-dbbk7\") pod \"ceilometer-0\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.621094 4966 scope.go:117] "RemoveContainer" containerID="0e28cef0856b356411a6187a94b70704373c7002299159e6be4c1ba2299eabb6" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.644577 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:47 crc kubenswrapper[4966]: I1217 08:41:47.692636 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.909090 4966 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/f31e26a6b7954078441163979a045c37db96d3a4c96c03c03c417fb7ba12e1cf/diff" to get inode usage: stat /var/lib/containers/storage/overlay/f31e26a6b7954078441163979a045c37db96d3a4c96c03c03c417fb7ba12e1cf/diff: no such file or directory, extraDiskErr: Dec 17 08:41:47 crc kubenswrapper[4966]: E1217 08:41:47.943984 4966 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/8a053375804b68bc0fc34af12bd4346c21c3069eadc5a5406abfd3824ca5cf28/diff" to get inode usage: stat /var/lib/containers/storage/overlay/8a053375804b68bc0fc34af12bd4346c21c3069eadc5a5406abfd3824ca5cf28/diff: no such file or directory, extraDiskErr: Dec 17 08:41:48 crc kubenswrapper[4966]: I1217 08:41:48.238582 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 17 08:41:48 crc kubenswrapper[4966]: I1217 08:41:48.243681 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"748a0d45-f457-4cdc-9c6b-7fb6a1d57ad0","Type":"ContainerStarted","Data":"be0afc022af08de77e41d29c7d0b37e3b1bf8042a8ebb3ee100fd82abe8d4659"} Dec 17 08:41:48 crc kubenswrapper[4966]: I1217 08:41:48.276658 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.276631527 podStartE2EDuration="4.276631527s" podCreationTimestamp="2025-12-17 08:41:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:41:48.276188204 +0000 UTC m=+1243.821258146" watchObservedRunningTime="2025-12-17 08:41:48.276631527 +0000 UTC m=+1243.821701479" Dec 17 08:41:48 crc kubenswrapper[4966]: I1217 08:41:48.327182 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:41:48 crc kubenswrapper[4966]: I1217 08:41:48.333264 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 08:41:48 crc kubenswrapper[4966]: E1217 08:41:48.349429 4966 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/54facdeda4f0ac2a9b8940abe1e23ed5ff8a057428c11dabfc9316ceb3219080/diff" to get inode usage: stat /var/lib/containers/storage/overlay/54facdeda4f0ac2a9b8940abe1e23ed5ff8a057428c11dabfc9316ceb3219080/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_heat-engine-b67cfd874-bklxp_f33826e0-0269-4b99-9dce-fcaf62f67bed/heat-engine/0.log" to get inode usage: stat /var/log/pods/openstack_heat-engine-b67cfd874-bklxp_f33826e0-0269-4b99-9dce-fcaf62f67bed/heat-engine/0.log: no such file or directory Dec 17 08:41:48 crc kubenswrapper[4966]: E1217 08:41:48.809109 4966 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/88c99c1404a4ea0577e6571a5a638ae40f162fa5a4b3ddf6a7391678a77316b0/diff" to get inode usage: stat /var/lib/containers/storage/overlay/88c99c1404a4ea0577e6571a5a638ae40f162fa5a4b3ddf6a7391678a77316b0/diff: no such file or directory, extraDiskErr: Dec 17 08:41:48 crc kubenswrapper[4966]: I1217 08:41:48.844286 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e49dad7-7eb2-43bb-afcf-068e4b517ab4" path="/var/lib/kubelet/pods/2e49dad7-7eb2-43bb-afcf-068e4b517ab4/volumes" Dec 17 08:41:49 crc kubenswrapper[4966]: I1217 08:41:49.258120 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dc23d16c-1ee3-4ee5-a102-e72e1571c032","Type":"ContainerStarted","Data":"11545f648b57acc6f26cb9857609dbc0593303ad1df512ab1c0e07a59708aec6"} Dec 17 08:41:49 crc kubenswrapper[4966]: I1217 08:41:49.258176 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dc23d16c-1ee3-4ee5-a102-e72e1571c032","Type":"ContainerStarted","Data":"d3f6a50be1cdb9a39695b3d79f4ac8a8ed0aaacc332751d333749f33b79f29b4"} Dec 17 08:41:49 crc kubenswrapper[4966]: I1217 08:41:49.258236 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:49 crc kubenswrapper[4966]: I1217 08:41:49.260937 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerStarted","Data":"42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457"} Dec 17 08:41:49 crc kubenswrapper[4966]: I1217 08:41:49.260971 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerStarted","Data":"8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea"} Dec 17 08:41:49 crc kubenswrapper[4966]: I1217 08:41:49.260980 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerStarted","Data":"05ca8991e4a5d89119bca17c9e7efd01bdca2eb5aec3d6b5d486d65f53f582a4"} Dec 17 08:41:49 crc kubenswrapper[4966]: I1217 08:41:49.276636 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.276615477 podStartE2EDuration="2.276615477s" podCreationTimestamp="2025-12-17 08:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:41:49.272601287 +0000 UTC m=+1244.817671229" watchObservedRunningTime="2025-12-17 08:41:49.276615477 +0000 UTC m=+1244.821685419" Dec 17 08:41:50 crc kubenswrapper[4966]: I1217 08:41:50.275390 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerStarted","Data":"7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25"} Dec 17 08:41:51 crc kubenswrapper[4966]: I1217 08:41:51.284861 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerStarted","Data":"2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3"} Dec 17 08:41:52 crc kubenswrapper[4966]: I1217 08:41:52.292912 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 17 08:41:52 crc kubenswrapper[4966]: I1217 08:41:52.389071 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6913403479999998 podStartE2EDuration="5.389052134s" podCreationTimestamp="2025-12-17 08:41:47 +0000 UTC" firstStartedPulling="2025-12-17 08:41:48.333033655 +0000 UTC m=+1243.878103597" lastFinishedPulling="2025-12-17 08:41:51.030745441 +0000 UTC m=+1246.575815383" observedRunningTime="2025-12-17 08:41:52.377638451 +0000 UTC m=+1247.922708403" watchObservedRunningTime="2025-12-17 08:41:52.389052134 +0000 UTC m=+1247.934122076" Dec 17 08:41:53 crc kubenswrapper[4966]: I1217 08:41:53.834179 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 17 08:41:53 crc kubenswrapper[4966]: I1217 08:41:53.834246 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 17 08:41:53 crc kubenswrapper[4966]: I1217 08:41:53.868814 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 17 08:41:53 crc kubenswrapper[4966]: I1217 08:41:53.880428 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 17 08:41:54 crc kubenswrapper[4966]: I1217 08:41:54.311534 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 17 08:41:54 crc kubenswrapper[4966]: I1217 08:41:54.311691 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 17 08:41:54 crc kubenswrapper[4966]: I1217 08:41:54.623582 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:54 crc kubenswrapper[4966]: I1217 08:41:54.623642 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:54 crc kubenswrapper[4966]: I1217 08:41:54.663476 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:54 crc kubenswrapper[4966]: I1217 08:41:54.710850 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.070638 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.153540 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data-custom\") pod \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.153749 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data\") pod \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.153865 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-combined-ca-bundle\") pod \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.154001 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q258b\" (UniqueName: \"kubernetes.io/projected/6c618545-5898-4e00-9ed3-5f34f2ad77c8-kube-api-access-q258b\") pod \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\" (UID: \"6c618545-5898-4e00-9ed3-5f34f2ad77c8\") " Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.182305 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6c618545-5898-4e00-9ed3-5f34f2ad77c8" (UID: "6c618545-5898-4e00-9ed3-5f34f2ad77c8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.182408 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c618545-5898-4e00-9ed3-5f34f2ad77c8-kube-api-access-q258b" (OuterVolumeSpecName: "kube-api-access-q258b") pod "6c618545-5898-4e00-9ed3-5f34f2ad77c8" (UID: "6c618545-5898-4e00-9ed3-5f34f2ad77c8"). InnerVolumeSpecName "kube-api-access-q258b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.220242 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c618545-5898-4e00-9ed3-5f34f2ad77c8" (UID: "6c618545-5898-4e00-9ed3-5f34f2ad77c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.238575 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data" (OuterVolumeSpecName: "config-data") pod "6c618545-5898-4e00-9ed3-5f34f2ad77c8" (UID: "6c618545-5898-4e00-9ed3-5f34f2ad77c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.257060 4966 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.257112 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.257125 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c618545-5898-4e00-9ed3-5f34f2ad77c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.257142 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q258b\" (UniqueName: \"kubernetes.io/projected/6c618545-5898-4e00-9ed3-5f34f2ad77c8-kube-api-access-q258b\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.320755 4966 generic.go:334] "Generic (PLEG): container finished" podID="6c618545-5898-4e00-9ed3-5f34f2ad77c8" containerID="d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275" exitCode=137 Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.320801 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.320849 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" event={"ID":"6c618545-5898-4e00-9ed3-5f34f2ad77c8","Type":"ContainerDied","Data":"d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275"} Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.320909 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7ccdc8f9cd-5gvvt" event={"ID":"6c618545-5898-4e00-9ed3-5f34f2ad77c8","Type":"ContainerDied","Data":"72f776d3e93779e4cbc17a76db405133174c02f728a50829e3bb4b99e93b517b"} Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.320934 4966 scope.go:117] "RemoveContainer" containerID="d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.321665 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.321699 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.350929 4966 scope.go:117] "RemoveContainer" containerID="d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275" Dec 17 08:41:55 crc kubenswrapper[4966]: E1217 08:41:55.359295 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275\": container with ID starting with d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275 not found: ID does not exist" containerID="d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.359519 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275"} err="failed to get container status \"d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275\": rpc error: code = NotFound desc = could not find container \"d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275\": container with ID starting with d6a849099dbd79d8002b9e864072477b8f02c696f2f785270e7591600e0fe275 not found: ID does not exist" Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.364201 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7ccdc8f9cd-5gvvt"] Dec 17 08:41:55 crc kubenswrapper[4966]: I1217 08:41:55.378785 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7ccdc8f9cd-5gvvt"] Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.800776 4966 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-conmon-f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-conmon-f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f.scope: no such file or directory Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.800843 4966 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-f005b6a86741d9fa292e0cfa1e691ecc22f06b1e493ac5482f406109e10c359f.scope: no such file or directory Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.800857 4966 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-conmon-854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-conmon-854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2.scope: no such file or directory Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.800885 4966 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-854370e2864bd949ede5f80f7623f216209d5d96fc86e18191c1fb2cce3d8fc2.scope: no such file or directory Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.802434 4966 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02f6a226_d49b_4726_8116_4b9ab2837338.slice/crio-conmon-21cb6a52d5dbe8f6273ea2d4a3330d298f871767fe3b4b5b28e6a463c09fef72.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02f6a226_d49b_4726_8116_4b9ab2837338.slice/crio-conmon-21cb6a52d5dbe8f6273ea2d4a3330d298f871767fe3b4b5b28e6a463c09fef72.scope: no such file or directory Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.802455 4966 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02f6a226_d49b_4726_8116_4b9ab2837338.slice/crio-21cb6a52d5dbe8f6273ea2d4a3330d298f871767fe3b4b5b28e6a463c09fef72.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02f6a226_d49b_4726_8116_4b9ab2837338.slice/crio-21cb6a52d5dbe8f6273ea2d4a3330d298f871767fe3b4b5b28e6a463c09fef72.scope: no such file or directory Dec 17 08:41:55 crc kubenswrapper[4966]: E1217 08:41:55.806179 4966 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3e02105_d411_4d80_9d37_ae44d4cd8a4c.slice/crio-0851c0256cf554c3732ecb7d2aa6d032f094cc221bb092bc3aec3214caf1e279: Error finding container 0851c0256cf554c3732ecb7d2aa6d032f094cc221bb092bc3aec3214caf1e279: Status 404 returned error can't find the container with id 0851c0256cf554c3732ecb7d2aa6d032f094cc221bb092bc3aec3214caf1e279 Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.807576 4966 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e49dad7_7eb2_43bb_afcf_068e4b517ab4.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e49dad7_7eb2_43bb_afcf_068e4b517ab4.slice: no such file or directory Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.810057 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-a524f479782fd6499342c20baa06f3f96b12a69ab2b235e9016cd3619a8be51f WatchSource:0}: Error finding container a524f479782fd6499342c20baa06f3f96b12a69ab2b235e9016cd3619a8be51f: Status 404 returned error can't find the container with id a524f479782fd6499342c20baa06f3f96b12a69ab2b235e9016cd3619a8be51f Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.810437 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983.scope WatchSource:0}: Error finding container ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983: Status 404 returned error can't find the container with id ce90c441203fd55cda73d94dfa3c66de9de0fed8bb918de12f1e7f44afc42983 Dec 17 08:41:55 crc kubenswrapper[4966]: W1217 08:41:55.811069 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6894341_8765_4100_a0d9_c2f8b40f9607.slice/crio-8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89.scope WatchSource:0}: Error finding container 8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89: Status 404 returned error can't find the container with id 8eee7dbae9eebd84cf5fcc1e05cf51f18465c3bd617e9fb0e5a670479100ad89 Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.172819 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.284998 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data-custom\") pod \"0b528e6e-b726-4895-910c-8544dcf26417\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.285060 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdmsv\" (UniqueName: \"kubernetes.io/projected/0b528e6e-b726-4895-910c-8544dcf26417-kube-api-access-vdmsv\") pod \"0b528e6e-b726-4895-910c-8544dcf26417\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.285093 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-combined-ca-bundle\") pod \"0b528e6e-b726-4895-910c-8544dcf26417\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.285217 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data\") pod \"0b528e6e-b726-4895-910c-8544dcf26417\" (UID: \"0b528e6e-b726-4895-910c-8544dcf26417\") " Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.289775 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0b528e6e-b726-4895-910c-8544dcf26417" (UID: "0b528e6e-b726-4895-910c-8544dcf26417"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.292568 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b528e6e-b726-4895-910c-8544dcf26417-kube-api-access-vdmsv" (OuterVolumeSpecName: "kube-api-access-vdmsv") pod "0b528e6e-b726-4895-910c-8544dcf26417" (UID: "0b528e6e-b726-4895-910c-8544dcf26417"). InnerVolumeSpecName "kube-api-access-vdmsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.337847 4966 generic.go:334] "Generic (PLEG): container finished" podID="0b528e6e-b726-4895-910c-8544dcf26417" containerID="e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c" exitCode=137 Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.338785 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6c76d785bf-vjmcl" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.339204 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6c76d785bf-vjmcl" event={"ID":"0b528e6e-b726-4895-910c-8544dcf26417","Type":"ContainerDied","Data":"e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c"} Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.339233 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6c76d785bf-vjmcl" event={"ID":"0b528e6e-b726-4895-910c-8544dcf26417","Type":"ContainerDied","Data":"17e8cfd2a7271403fecef231c419eec9efda78d56d809247e0310b9c0b75cbdc"} Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.339251 4966 scope.go:117] "RemoveContainer" containerID="e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.339981 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b528e6e-b726-4895-910c-8544dcf26417" (UID: "0b528e6e-b726-4895-910c-8544dcf26417"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.345895 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data" (OuterVolumeSpecName: "config-data") pod "0b528e6e-b726-4895-910c-8544dcf26417" (UID: "0b528e6e-b726-4895-910c-8544dcf26417"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.376040 4966 scope.go:117] "RemoveContainer" containerID="e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c" Dec 17 08:41:56 crc kubenswrapper[4966]: E1217 08:41:56.377332 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c\": container with ID starting with e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c not found: ID does not exist" containerID="e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.377386 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c"} err="failed to get container status \"e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c\": rpc error: code = NotFound desc = could not find container \"e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c\": container with ID starting with e326d9a406d6cad673d4ff4178a5925280a48a89fc5fda4ede22219c52129e8c not found: ID does not exist" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.387031 4966 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.387070 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdmsv\" (UniqueName: \"kubernetes.io/projected/0b528e6e-b726-4895-910c-8544dcf26417-kube-api-access-vdmsv\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.387083 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.387096 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b528e6e-b726-4895-910c-8544dcf26417-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.685904 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6c76d785bf-vjmcl"] Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.750193 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6c76d785bf-vjmcl"] Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.852040 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b528e6e-b726-4895-910c-8544dcf26417" path="/var/lib/kubelet/pods/0b528e6e-b726-4895-910c-8544dcf26417/volumes" Dec 17 08:41:56 crc kubenswrapper[4966]: I1217 08:41:56.853004 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c618545-5898-4e00-9ed3-5f34f2ad77c8" path="/var/lib/kubelet/pods/6c618545-5898-4e00-9ed3-5f34f2ad77c8/volumes" Dec 17 08:41:57 crc kubenswrapper[4966]: I1217 08:41:57.230245 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 17 08:41:57 crc kubenswrapper[4966]: I1217 08:41:57.230588 4966 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 17 08:41:57 crc kubenswrapper[4966]: I1217 08:41:57.237156 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 17 08:41:57 crc kubenswrapper[4966]: I1217 08:41:57.616935 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:57 crc kubenswrapper[4966]: I1217 08:41:57.617071 4966 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 17 08:41:57 crc kubenswrapper[4966]: I1217 08:41:57.822533 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 17 08:41:57 crc kubenswrapper[4966]: I1217 08:41:57.836088 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.387402 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-7bzdn"] Dec 17 08:41:58 crc kubenswrapper[4966]: E1217 08:41:58.388096 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b528e6e-b726-4895-910c-8544dcf26417" containerName="heat-api" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.388111 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b528e6e-b726-4895-910c-8544dcf26417" containerName="heat-api" Dec 17 08:41:58 crc kubenswrapper[4966]: E1217 08:41:58.388137 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c618545-5898-4e00-9ed3-5f34f2ad77c8" containerName="heat-cfnapi" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.388143 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c618545-5898-4e00-9ed3-5f34f2ad77c8" containerName="heat-cfnapi" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.388344 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b528e6e-b726-4895-910c-8544dcf26417" containerName="heat-api" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.388366 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c618545-5898-4e00-9ed3-5f34f2ad77c8" containerName="heat-cfnapi" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.388976 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.395404 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.397083 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.417670 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-7bzdn"] Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.434240 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz6zg\" (UniqueName: \"kubernetes.io/projected/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-kube-api-access-qz6zg\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.434308 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.434379 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-config-data\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.434551 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-scripts\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.530176 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.531685 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.534456 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.538104 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-scripts\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.538210 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz6zg\" (UniqueName: \"kubernetes.io/projected/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-kube-api-access-qz6zg\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.538245 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.538289 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-config-data\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.549343 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-config-data\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.560337 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.571366 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-scripts\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.574010 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.639238 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz6zg\" (UniqueName: \"kubernetes.io/projected/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-kube-api-access-qz6zg\") pod \"nova-cell0-cell-mapping-7bzdn\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.640143 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-config-data\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.640211 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-logs\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.648080 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj458\" (UniqueName: \"kubernetes.io/projected/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-kube-api-access-pj458\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.648311 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.704556 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.704808 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.706270 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.716108 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.750407 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-config-data\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.750474 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-logs\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.750545 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj458\" (UniqueName: \"kubernetes.io/projected/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-kube-api-access-pj458\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.750573 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.750666 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-config-data\") pod \"nova-scheduler-0\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.750719 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.750818 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5qxw\" (UniqueName: \"kubernetes.io/projected/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-kube-api-access-q5qxw\") pod \"nova-scheduler-0\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.756069 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.765774 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-config-data\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.770714 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-logs\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.816121 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.855747 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj458\" (UniqueName: \"kubernetes.io/projected/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-kube-api-access-pj458\") pod \"nova-api-0\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " pod="openstack/nova-api-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.860117 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-config-data\") pod \"nova-scheduler-0\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.860252 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5qxw\" (UniqueName: \"kubernetes.io/projected/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-kube-api-access-q5qxw\") pod \"nova-scheduler-0\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.860372 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.900736 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-config-data\") pod \"nova-scheduler-0\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.901314 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.934415 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.951545 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5qxw\" (UniqueName: \"kubernetes.io/projected/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-kube-api-access-q5qxw\") pod \"nova-scheduler-0\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " pod="openstack/nova-scheduler-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.959019 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.971485 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2091b27-0810-4703-b5fc-389a9de073b9-logs\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.971541 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7k9x\" (UniqueName: \"kubernetes.io/projected/a2091b27-0810-4703-b5fc-389a9de073b9-kube-api-access-s7k9x\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.971737 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-config-data\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.971796 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:58 crc kubenswrapper[4966]: I1217 08:41:58.977375 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.013800 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.028172 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.072191 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.074364 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.104078 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.106502 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2091b27-0810-4703-b5fc-389a9de073b9-logs\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.106537 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7k9x\" (UniqueName: \"kubernetes.io/projected/a2091b27-0810-4703-b5fc-389a9de073b9-kube-api-access-s7k9x\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.106592 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-config-data\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.106626 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.112926 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.115453 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2091b27-0810-4703-b5fc-389a9de073b9-logs\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.145541 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7k9x\" (UniqueName: \"kubernetes.io/projected/a2091b27-0810-4703-b5fc-389a9de073b9-kube-api-access-s7k9x\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.150452 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.167187 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-config-data\") pod \"nova-metadata-0\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " pod="openstack/nova-metadata-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.241788 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htn8h\" (UniqueName: \"kubernetes.io/projected/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-kube-api-access-htn8h\") pod \"nova-cell1-novncproxy-0\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.242155 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.242220 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.246001 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.258675 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d88fdb6c9-zmf26"] Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.262139 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.288436 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d88fdb6c9-zmf26"] Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.316683 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.343725 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htn8h\" (UniqueName: \"kubernetes.io/projected/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-kube-api-access-htn8h\") pod \"nova-cell1-novncproxy-0\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.343773 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.343826 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-sb\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.343858 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.343924 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-config\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.343972 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-swift-storage-0\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.343998 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-nb\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.344105 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4pds\" (UniqueName: \"kubernetes.io/projected/06f02ab9-5bb8-47ba-88a8-9971cf378afa-kube-api-access-t4pds\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.344151 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-svc\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.357254 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.367906 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.380190 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htn8h\" (UniqueName: \"kubernetes.io/projected/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-kube-api-access-htn8h\") pod \"nova-cell1-novncproxy-0\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.447144 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-swift-storage-0\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.447188 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-nb\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.447268 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4pds\" (UniqueName: \"kubernetes.io/projected/06f02ab9-5bb8-47ba-88a8-9971cf378afa-kube-api-access-t4pds\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.447303 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-svc\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.447353 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-sb\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.447399 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-config\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.448246 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-config\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.451130 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-svc\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.451681 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-sb\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.452527 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-nb\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.452742 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-swift-storage-0\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.461180 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.469315 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4pds\" (UniqueName: \"kubernetes.io/projected/06f02ab9-5bb8-47ba-88a8-9971cf378afa-kube-api-access-t4pds\") pod \"dnsmasq-dns-5d88fdb6c9-zmf26\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.584191 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.641447 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-7bzdn"] Dec 17 08:41:59 crc kubenswrapper[4966]: I1217 08:41:59.858715 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.170460 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:00 crc kubenswrapper[4966]: W1217 08:42:00.192510 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod248f6a8e_1c9c_42e7_9e16_9f16aa6049f8.slice/crio-d757f6a7640c90e29fa36db8c5b81ae51adb9c1adfe54b0ea0b5cdeee9ece6cc WatchSource:0}: Error finding container d757f6a7640c90e29fa36db8c5b81ae51adb9c1adfe54b0ea0b5cdeee9ece6cc: Status 404 returned error can't find the container with id d757f6a7640c90e29fa36db8c5b81ae51adb9c1adfe54b0ea0b5cdeee9ece6cc Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.254675 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.283037 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.295186 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d88fdb6c9-zmf26"] Dec 17 08:42:00 crc kubenswrapper[4966]: W1217 08:42:00.295280 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2091b27_0810_4703_b5fc_389a9de073b9.slice/crio-6f66b3a981c2cdecc798fcb062ae6e1fabd805b361b34b581321f3ec6567d327 WatchSource:0}: Error finding container 6f66b3a981c2cdecc798fcb062ae6e1fabd805b361b34b581321f3ec6567d327: Status 404 returned error can't find the container with id 6f66b3a981c2cdecc798fcb062ae6e1fabd805b361b34b581321f3ec6567d327 Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.412185 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7bzdn" event={"ID":"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa","Type":"ContainerStarted","Data":"e4ef0f804df1958cad65b3e71ba5895646044cf889412cd1913439c97da4c73f"} Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.412475 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7bzdn" event={"ID":"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa","Type":"ContainerStarted","Data":"35aa72821012d5af9a696c15c60c8d65d4e5d24c1615a9d39fe2ce37d5a88e65"} Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.416630 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4a5ceb27-fa91-498b-90d6-1828d9efbf5f","Type":"ContainerStarted","Data":"d1459b38e7df3398f23791cad7435efa32e68fe980aa50619f343b5c68d7bc00"} Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.429466 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" event={"ID":"06f02ab9-5bb8-47ba-88a8-9971cf378afa","Type":"ContainerStarted","Data":"11e9e33eddd638a04f5ea126c1248301b5b2a3eb572fd1553f33584ca1e46264"} Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.439163 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a2091b27-0810-4703-b5fc-389a9de073b9","Type":"ContainerStarted","Data":"6f66b3a981c2cdecc798fcb062ae6e1fabd805b361b34b581321f3ec6567d327"} Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.442798 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-7bzdn" podStartSLOduration=2.442785731 podStartE2EDuration="2.442785731s" podCreationTimestamp="2025-12-17 08:41:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:00.427409519 +0000 UTC m=+1255.972479461" watchObservedRunningTime="2025-12-17 08:42:00.442785731 +0000 UTC m=+1255.987855673" Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.442949 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3c2e3f9-8b0b-4fb6-8404-035c11650e41","Type":"ContainerStarted","Data":"2e52d97b0e92ae614d8dead8675ca6e39d275cea344c8c789ab851b697238063"} Dec 17 08:42:00 crc kubenswrapper[4966]: I1217 08:42:00.453439 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8","Type":"ContainerStarted","Data":"d757f6a7640c90e29fa36db8c5b81ae51adb9c1adfe54b0ea0b5cdeee9ece6cc"} Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.173643 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tdhl5"] Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.175127 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.181000 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.181086 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.189436 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tdhl5"] Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.286996 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-scripts\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.287357 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rmtv\" (UniqueName: \"kubernetes.io/projected/a272c7ab-dc3f-49af-9f02-6fe4b552c217-kube-api-access-2rmtv\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.287442 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.287536 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-config-data\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.388999 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.389090 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-config-data\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.389159 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-scripts\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.389206 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rmtv\" (UniqueName: \"kubernetes.io/projected/a272c7ab-dc3f-49af-9f02-6fe4b552c217-kube-api-access-2rmtv\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.393586 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-scripts\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.394400 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-config-data\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.395943 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.417422 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rmtv\" (UniqueName: \"kubernetes.io/projected/a272c7ab-dc3f-49af-9f02-6fe4b552c217-kube-api-access-2rmtv\") pod \"nova-cell1-conductor-db-sync-tdhl5\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.480401 4966 generic.go:334] "Generic (PLEG): container finished" podID="06f02ab9-5bb8-47ba-88a8-9971cf378afa" containerID="99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51" exitCode=0 Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.482092 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" event={"ID":"06f02ab9-5bb8-47ba-88a8-9971cf378afa","Type":"ContainerDied","Data":"99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51"} Dec 17 08:42:01 crc kubenswrapper[4966]: I1217 08:42:01.501853 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:02 crc kubenswrapper[4966]: I1217 08:42:02.579415 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 17 08:42:02 crc kubenswrapper[4966]: I1217 08:42:02.597935 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:04 crc kubenswrapper[4966]: I1217 08:42:04.790647 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tdhl5"] Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.403000 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.403561 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="ceilometer-central-agent" containerID="cri-o://8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea" gracePeriod=30 Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.403696 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="ceilometer-notification-agent" containerID="cri-o://42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457" gracePeriod=30 Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.403708 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="sg-core" containerID="cri-o://7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25" gracePeriod=30 Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.403818 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="proxy-httpd" containerID="cri-o://2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3" gracePeriod=30 Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.413276 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.200:3000/\": EOF" Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.554646 4966 generic.go:334] "Generic (PLEG): container finished" podID="c0f67424-421d-4dc9-baad-af7ad452b801" containerID="7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25" exitCode=2 Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.554730 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerDied","Data":"7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.556351 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4a5ceb27-fa91-498b-90d6-1828d9efbf5f","Type":"ContainerStarted","Data":"60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.556476 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="4a5ceb27-fa91-498b-90d6-1828d9efbf5f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a" gracePeriod=30 Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.559955 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" event={"ID":"06f02ab9-5bb8-47ba-88a8-9971cf378afa","Type":"ContainerStarted","Data":"71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.560083 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.563642 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a2091b27-0810-4703-b5fc-389a9de073b9","Type":"ContainerStarted","Data":"647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.563692 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a2091b27-0810-4703-b5fc-389a9de073b9","Type":"ContainerStarted","Data":"7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.563712 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a2091b27-0810-4703-b5fc-389a9de073b9" containerName="nova-metadata-log" containerID="cri-o://7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48" gracePeriod=30 Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.563778 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a2091b27-0810-4703-b5fc-389a9de073b9" containerName="nova-metadata-metadata" containerID="cri-o://647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae" gracePeriod=30 Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.568029 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3c2e3f9-8b0b-4fb6-8404-035c11650e41","Type":"ContainerStarted","Data":"cbfcd389f3c7812c50c33f46a05c3b30185207d043879432c77df1ab09c47d21"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.570574 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8","Type":"ContainerStarted","Data":"a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.570602 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8","Type":"ContainerStarted","Data":"aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.572624 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tdhl5" event={"ID":"a272c7ab-dc3f-49af-9f02-6fe4b552c217","Type":"ContainerStarted","Data":"4eb94a4fb5c18d6acd037c4294a3352d575e3b79ac2e5f740d46fa74aa105530"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.572670 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tdhl5" event={"ID":"a272c7ab-dc3f-49af-9f02-6fe4b552c217","Type":"ContainerStarted","Data":"17dee0eec6af7b7cb23582eac74d110f00449db17cbdce27fb78f84bd2c0f8d9"} Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.610553 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.6739346250000002 podStartE2EDuration="7.610532477s" podCreationTimestamp="2025-12-17 08:41:58 +0000 UTC" firstStartedPulling="2025-12-17 08:42:00.278934075 +0000 UTC m=+1255.824004017" lastFinishedPulling="2025-12-17 08:42:04.215531927 +0000 UTC m=+1259.760601869" observedRunningTime="2025-12-17 08:42:05.571843645 +0000 UTC m=+1261.116913587" watchObservedRunningTime="2025-12-17 08:42:05.610532477 +0000 UTC m=+1261.155602419" Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.611731 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.695902478 podStartE2EDuration="7.61172668s" podCreationTimestamp="2025-12-17 08:41:58 +0000 UTC" firstStartedPulling="2025-12-17 08:42:00.300018374 +0000 UTC m=+1255.845088316" lastFinishedPulling="2025-12-17 08:42:04.215842576 +0000 UTC m=+1259.760912518" observedRunningTime="2025-12-17 08:42:05.604252354 +0000 UTC m=+1261.149322296" watchObservedRunningTime="2025-12-17 08:42:05.61172668 +0000 UTC m=+1261.156796622" Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.665163 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.3282338879999998 podStartE2EDuration="7.665143905s" podCreationTimestamp="2025-12-17 08:41:58 +0000 UTC" firstStartedPulling="2025-12-17 08:41:59.877488359 +0000 UTC m=+1255.422558301" lastFinishedPulling="2025-12-17 08:42:04.214398376 +0000 UTC m=+1259.759468318" observedRunningTime="2025-12-17 08:42:05.632286113 +0000 UTC m=+1261.177356065" watchObservedRunningTime="2025-12-17 08:42:05.665143905 +0000 UTC m=+1261.210213847" Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.708366 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.690012097 podStartE2EDuration="7.708345101s" podCreationTimestamp="2025-12-17 08:41:58 +0000 UTC" firstStartedPulling="2025-12-17 08:42:00.195966099 +0000 UTC m=+1255.741036041" lastFinishedPulling="2025-12-17 08:42:04.214299103 +0000 UTC m=+1259.759369045" observedRunningTime="2025-12-17 08:42:05.68720568 +0000 UTC m=+1261.232275632" watchObservedRunningTime="2025-12-17 08:42:05.708345101 +0000 UTC m=+1261.253415043" Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.720977 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" podStartSLOduration=6.720956107 podStartE2EDuration="6.720956107s" podCreationTimestamp="2025-12-17 08:41:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:05.713324557 +0000 UTC m=+1261.258394509" watchObservedRunningTime="2025-12-17 08:42:05.720956107 +0000 UTC m=+1261.266026049" Dec 17 08:42:05 crc kubenswrapper[4966]: I1217 08:42:05.753926 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-tdhl5" podStartSLOduration=4.75390015 podStartE2EDuration="4.75390015s" podCreationTimestamp="2025-12-17 08:42:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:05.736521134 +0000 UTC m=+1261.281591066" watchObservedRunningTime="2025-12-17 08:42:05.75390015 +0000 UTC m=+1261.298970102" Dec 17 08:42:06 crc kubenswrapper[4966]: E1217 08:42:06.055088 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0f67424_421d_4dc9_baad_af7ad452b801.slice/crio-conmon-8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea.scope\": RecentStats: unable to find data in memory cache]" Dec 17 08:42:06 crc kubenswrapper[4966]: I1217 08:42:06.584147 4966 generic.go:334] "Generic (PLEG): container finished" podID="c0f67424-421d-4dc9-baad-af7ad452b801" containerID="2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3" exitCode=0 Dec 17 08:42:06 crc kubenswrapper[4966]: I1217 08:42:06.584424 4966 generic.go:334] "Generic (PLEG): container finished" podID="c0f67424-421d-4dc9-baad-af7ad452b801" containerID="8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea" exitCode=0 Dec 17 08:42:06 crc kubenswrapper[4966]: I1217 08:42:06.584226 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerDied","Data":"2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3"} Dec 17 08:42:06 crc kubenswrapper[4966]: I1217 08:42:06.584503 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerDied","Data":"8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea"} Dec 17 08:42:06 crc kubenswrapper[4966]: I1217 08:42:06.586608 4966 generic.go:334] "Generic (PLEG): container finished" podID="a2091b27-0810-4703-b5fc-389a9de073b9" containerID="7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48" exitCode=143 Dec 17 08:42:06 crc kubenswrapper[4966]: I1217 08:42:06.586704 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a2091b27-0810-4703-b5fc-389a9de073b9","Type":"ContainerDied","Data":"7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48"} Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.469944 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.548644 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-config-data\") pod \"a2091b27-0810-4703-b5fc-389a9de073b9\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.548696 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-combined-ca-bundle\") pod \"a2091b27-0810-4703-b5fc-389a9de073b9\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.548771 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2091b27-0810-4703-b5fc-389a9de073b9-logs\") pod \"a2091b27-0810-4703-b5fc-389a9de073b9\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.548895 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7k9x\" (UniqueName: \"kubernetes.io/projected/a2091b27-0810-4703-b5fc-389a9de073b9-kube-api-access-s7k9x\") pod \"a2091b27-0810-4703-b5fc-389a9de073b9\" (UID: \"a2091b27-0810-4703-b5fc-389a9de073b9\") " Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.549663 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2091b27-0810-4703-b5fc-389a9de073b9-logs" (OuterVolumeSpecName: "logs") pod "a2091b27-0810-4703-b5fc-389a9de073b9" (UID: "a2091b27-0810-4703-b5fc-389a9de073b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.556041 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2091b27-0810-4703-b5fc-389a9de073b9-kube-api-access-s7k9x" (OuterVolumeSpecName: "kube-api-access-s7k9x") pod "a2091b27-0810-4703-b5fc-389a9de073b9" (UID: "a2091b27-0810-4703-b5fc-389a9de073b9"). InnerVolumeSpecName "kube-api-access-s7k9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.588146 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2091b27-0810-4703-b5fc-389a9de073b9" (UID: "a2091b27-0810-4703-b5fc-389a9de073b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.597154 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-config-data" (OuterVolumeSpecName: "config-data") pod "a2091b27-0810-4703-b5fc-389a9de073b9" (UID: "a2091b27-0810-4703-b5fc-389a9de073b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.616662 4966 generic.go:334] "Generic (PLEG): container finished" podID="a2091b27-0810-4703-b5fc-389a9de073b9" containerID="647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae" exitCode=0 Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.616713 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a2091b27-0810-4703-b5fc-389a9de073b9","Type":"ContainerDied","Data":"647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae"} Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.616739 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a2091b27-0810-4703-b5fc-389a9de073b9","Type":"ContainerDied","Data":"6f66b3a981c2cdecc798fcb062ae6e1fabd805b361b34b581321f3ec6567d327"} Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.616754 4966 scope.go:117] "RemoveContainer" containerID="647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.616909 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.651441 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.651487 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2091b27-0810-4703-b5fc-389a9de073b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.651505 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2091b27-0810-4703-b5fc-389a9de073b9-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.651517 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7k9x\" (UniqueName: \"kubernetes.io/projected/a2091b27-0810-4703-b5fc-389a9de073b9-kube-api-access-s7k9x\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.695314 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.697058 4966 scope.go:117] "RemoveContainer" containerID="7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.719136 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.728998 4966 scope.go:117] "RemoveContainer" containerID="647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae" Dec 17 08:42:07 crc kubenswrapper[4966]: E1217 08:42:07.731291 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae\": container with ID starting with 647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae not found: ID does not exist" containerID="647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.731322 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae"} err="failed to get container status \"647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae\": rpc error: code = NotFound desc = could not find container \"647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae\": container with ID starting with 647a389c13a03e6f5dc2b0516b1f34f2fa2d2dfd7b04088781f23679155726ae not found: ID does not exist" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.731344 4966 scope.go:117] "RemoveContainer" containerID="7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.731395 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:07 crc kubenswrapper[4966]: E1217 08:42:07.731740 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2091b27-0810-4703-b5fc-389a9de073b9" containerName="nova-metadata-metadata" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.731755 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2091b27-0810-4703-b5fc-389a9de073b9" containerName="nova-metadata-metadata" Dec 17 08:42:07 crc kubenswrapper[4966]: E1217 08:42:07.731792 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2091b27-0810-4703-b5fc-389a9de073b9" containerName="nova-metadata-log" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.731798 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2091b27-0810-4703-b5fc-389a9de073b9" containerName="nova-metadata-log" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.731997 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2091b27-0810-4703-b5fc-389a9de073b9" containerName="nova-metadata-log" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.732013 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2091b27-0810-4703-b5fc-389a9de073b9" containerName="nova-metadata-metadata" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.732975 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: E1217 08:42:07.733815 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48\": container with ID starting with 7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48 not found: ID does not exist" containerID="7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.733839 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48"} err="failed to get container status \"7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48\": rpc error: code = NotFound desc = could not find container \"7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48\": container with ID starting with 7756a3d56e26d74b0d05774406fb6ee032cbbfd22eb5ab009bb7b3e5d9176c48 not found: ID does not exist" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.740258 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.740449 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.740450 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.868806 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.869508 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-logs\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.869644 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw8ck\" (UniqueName: \"kubernetes.io/projected/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-kube-api-access-vw8ck\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.869754 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.869905 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-config-data\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.972103 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.973116 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-logs\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.973433 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-logs\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.973156 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw8ck\" (UniqueName: \"kubernetes.io/projected/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-kube-api-access-vw8ck\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.973518 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.973581 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-config-data\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.976476 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.977112 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.979449 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-config-data\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:07 crc kubenswrapper[4966]: I1217 08:42:07.994185 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw8ck\" (UniqueName: \"kubernetes.io/projected/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-kube-api-access-vw8ck\") pod \"nova-metadata-0\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " pod="openstack/nova-metadata-0" Dec 17 08:42:08 crc kubenswrapper[4966]: I1217 08:42:08.090699 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:42:08 crc kubenswrapper[4966]: W1217 08:42:08.596032 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b99d0d7_30a1_4d05_996a_38a64e1e2bfd.slice/crio-6da804fdcee0f8c1c554b946537375c761e2482ec16cad6838f5bea4cac570c5 WatchSource:0}: Error finding container 6da804fdcee0f8c1c554b946537375c761e2482ec16cad6838f5bea4cac570c5: Status 404 returned error can't find the container with id 6da804fdcee0f8c1c554b946537375c761e2482ec16cad6838f5bea4cac570c5 Dec 17 08:42:08 crc kubenswrapper[4966]: I1217 08:42:08.597907 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:08 crc kubenswrapper[4966]: I1217 08:42:08.641216 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd","Type":"ContainerStarted","Data":"6da804fdcee0f8c1c554b946537375c761e2482ec16cad6838f5bea4cac570c5"} Dec 17 08:42:08 crc kubenswrapper[4966]: I1217 08:42:08.848045 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2091b27-0810-4703-b5fc-389a9de073b9" path="/var/lib/kubelet/pods/a2091b27-0810-4703-b5fc-389a9de073b9/volumes" Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.031294 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.031354 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.105682 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.105733 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.147312 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.463122 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.654753 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd","Type":"ContainerStarted","Data":"97b8c3fecbf1e56e636448b1b4043640278344250085acc71a316af6ef9e2e3d"} Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.654818 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd","Type":"ContainerStarted","Data":"cb463f8fbc0052ea2ac3e54a53fbd74c5772b2c07972319fd411fcba2eb701f2"} Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.694368 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 17 08:42:09 crc kubenswrapper[4966]: I1217 08:42:09.719246 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.719221961 podStartE2EDuration="2.719221961s" podCreationTimestamp="2025-12-17 08:42:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:09.684125108 +0000 UTC m=+1265.229195060" watchObservedRunningTime="2025-12-17 08:42:09.719221961 +0000 UTC m=+1265.264291903" Dec 17 08:42:10 crc kubenswrapper[4966]: I1217 08:42:10.113115 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 08:42:10 crc kubenswrapper[4966]: I1217 08:42:10.113142 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 08:42:10 crc kubenswrapper[4966]: I1217 08:42:10.668206 4966 generic.go:334] "Generic (PLEG): container finished" podID="072c5a1a-6292-4420-b5a1-dcd5dd93c5aa" containerID="e4ef0f804df1958cad65b3e71ba5895646044cf889412cd1913439c97da4c73f" exitCode=0 Dec 17 08:42:10 crc kubenswrapper[4966]: I1217 08:42:10.668280 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7bzdn" event={"ID":"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa","Type":"ContainerDied","Data":"e4ef0f804df1958cad65b3e71ba5895646044cf889412cd1913439c97da4c73f"} Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.153487 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.246749 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-combined-ca-bundle\") pod \"c0f67424-421d-4dc9-baad-af7ad452b801\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.246828 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-sg-core-conf-yaml\") pod \"c0f67424-421d-4dc9-baad-af7ad452b801\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.246926 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-log-httpd\") pod \"c0f67424-421d-4dc9-baad-af7ad452b801\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.246984 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-config-data\") pod \"c0f67424-421d-4dc9-baad-af7ad452b801\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.247006 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbbk7\" (UniqueName: \"kubernetes.io/projected/c0f67424-421d-4dc9-baad-af7ad452b801-kube-api-access-dbbk7\") pod \"c0f67424-421d-4dc9-baad-af7ad452b801\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.247072 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-scripts\") pod \"c0f67424-421d-4dc9-baad-af7ad452b801\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.247128 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-run-httpd\") pod \"c0f67424-421d-4dc9-baad-af7ad452b801\" (UID: \"c0f67424-421d-4dc9-baad-af7ad452b801\") " Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.247364 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c0f67424-421d-4dc9-baad-af7ad452b801" (UID: "c0f67424-421d-4dc9-baad-af7ad452b801"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.247806 4966 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.248046 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c0f67424-421d-4dc9-baad-af7ad452b801" (UID: "c0f67424-421d-4dc9-baad-af7ad452b801"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.269053 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0f67424-421d-4dc9-baad-af7ad452b801-kube-api-access-dbbk7" (OuterVolumeSpecName: "kube-api-access-dbbk7") pod "c0f67424-421d-4dc9-baad-af7ad452b801" (UID: "c0f67424-421d-4dc9-baad-af7ad452b801"). InnerVolumeSpecName "kube-api-access-dbbk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.270983 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-scripts" (OuterVolumeSpecName: "scripts") pod "c0f67424-421d-4dc9-baad-af7ad452b801" (UID: "c0f67424-421d-4dc9-baad-af7ad452b801"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.286533 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c0f67424-421d-4dc9-baad-af7ad452b801" (UID: "c0f67424-421d-4dc9-baad-af7ad452b801"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.349098 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.349125 4966 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0f67424-421d-4dc9-baad-af7ad452b801-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.349133 4966 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.349142 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbbk7\" (UniqueName: \"kubernetes.io/projected/c0f67424-421d-4dc9-baad-af7ad452b801-kube-api-access-dbbk7\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.363359 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-config-data" (OuterVolumeSpecName: "config-data") pod "c0f67424-421d-4dc9-baad-af7ad452b801" (UID: "c0f67424-421d-4dc9-baad-af7ad452b801"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.363945 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0f67424-421d-4dc9-baad-af7ad452b801" (UID: "c0f67424-421d-4dc9-baad-af7ad452b801"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.451443 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.451482 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0f67424-421d-4dc9-baad-af7ad452b801-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.688895 4966 generic.go:334] "Generic (PLEG): container finished" podID="c0f67424-421d-4dc9-baad-af7ad452b801" containerID="42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457" exitCode=0 Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.688985 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerDied","Data":"42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457"} Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.689039 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0f67424-421d-4dc9-baad-af7ad452b801","Type":"ContainerDied","Data":"05ca8991e4a5d89119bca17c9e7efd01bdca2eb5aec3d6b5d486d65f53f582a4"} Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.689025 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.689062 4966 scope.go:117] "RemoveContainer" containerID="2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.730357 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.739080 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.752757 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:11 crc kubenswrapper[4966]: E1217 08:42:11.755968 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="ceilometer-notification-agent" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.756004 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="ceilometer-notification-agent" Dec 17 08:42:11 crc kubenswrapper[4966]: E1217 08:42:11.756034 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="proxy-httpd" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.756041 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="proxy-httpd" Dec 17 08:42:11 crc kubenswrapper[4966]: E1217 08:42:11.756061 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="sg-core" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.756067 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="sg-core" Dec 17 08:42:11 crc kubenswrapper[4966]: E1217 08:42:11.756084 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="ceilometer-central-agent" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.756090 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="ceilometer-central-agent" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.756262 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="ceilometer-central-agent" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.756277 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="proxy-httpd" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.756292 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="sg-core" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.756307 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" containerName="ceilometer-notification-agent" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.758283 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.769143 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.769472 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.777370 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.808033 4966 scope.go:117] "RemoveContainer" containerID="7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.855302 4966 scope.go:117] "RemoveContainer" containerID="42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.864392 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-run-httpd\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.864461 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwfq9\" (UniqueName: \"kubernetes.io/projected/d4f6238b-6948-43ae-8a0e-1279ec490a71-kube-api-access-qwfq9\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.864487 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.864607 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-scripts\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.864644 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.864668 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-config-data\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.864710 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-log-httpd\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.916069 4966 scope.go:117] "RemoveContainer" containerID="8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.944842 4966 scope.go:117] "RemoveContainer" containerID="2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3" Dec 17 08:42:11 crc kubenswrapper[4966]: E1217 08:42:11.945443 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3\": container with ID starting with 2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3 not found: ID does not exist" containerID="2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.945503 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3"} err="failed to get container status \"2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3\": rpc error: code = NotFound desc = could not find container \"2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3\": container with ID starting with 2d9e9a49be98312c1c127f30c199d24eb2e3e6614a46456b118381782a684ef3 not found: ID does not exist" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.945536 4966 scope.go:117] "RemoveContainer" containerID="7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25" Dec 17 08:42:11 crc kubenswrapper[4966]: E1217 08:42:11.946060 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25\": container with ID starting with 7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25 not found: ID does not exist" containerID="7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.946095 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25"} err="failed to get container status \"7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25\": rpc error: code = NotFound desc = could not find container \"7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25\": container with ID starting with 7e00ac4db52ebe69927c22e4db0e168b34f8b38620c95c4619e032104b775c25 not found: ID does not exist" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.946117 4966 scope.go:117] "RemoveContainer" containerID="42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457" Dec 17 08:42:11 crc kubenswrapper[4966]: E1217 08:42:11.946403 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457\": container with ID starting with 42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457 not found: ID does not exist" containerID="42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.946441 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457"} err="failed to get container status \"42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457\": rpc error: code = NotFound desc = could not find container \"42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457\": container with ID starting with 42160c87d3707825a9e868e3e763116cdf806c7b38f500c1f7a3472770d5c457 not found: ID does not exist" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.946458 4966 scope.go:117] "RemoveContainer" containerID="8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea" Dec 17 08:42:11 crc kubenswrapper[4966]: E1217 08:42:11.947037 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea\": container with ID starting with 8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea not found: ID does not exist" containerID="8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.947087 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea"} err="failed to get container status \"8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea\": rpc error: code = NotFound desc = could not find container \"8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea\": container with ID starting with 8cfd25ee4b52bae53b88a4df5b437e35fa2ba13e2f183347fe7e65a9ce40d4ea not found: ID does not exist" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.974242 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-scripts\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.974294 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.974583 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-config-data\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.974693 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-log-httpd\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.974951 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-run-httpd\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.975036 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwfq9\" (UniqueName: \"kubernetes.io/projected/d4f6238b-6948-43ae-8a0e-1279ec490a71-kube-api-access-qwfq9\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.975060 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.989165 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-config-data\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.989847 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.975572 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-log-httpd\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:11 crc kubenswrapper[4966]: I1217 08:42:11.991691 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-run-httpd\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.002942 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.007209 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-scripts\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.058743 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwfq9\" (UniqueName: \"kubernetes.io/projected/d4f6238b-6948-43ae-8a0e-1279ec490a71-kube-api-access-qwfq9\") pod \"ceilometer-0\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " pod="openstack/ceilometer-0" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.167653 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.175781 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.291155 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qz6zg\" (UniqueName: \"kubernetes.io/projected/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-kube-api-access-qz6zg\") pod \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.291190 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-config-data\") pod \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.291380 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-scripts\") pod \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.291887 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-combined-ca-bundle\") pod \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\" (UID: \"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa\") " Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.296053 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-scripts" (OuterVolumeSpecName: "scripts") pod "072c5a1a-6292-4420-b5a1-dcd5dd93c5aa" (UID: "072c5a1a-6292-4420-b5a1-dcd5dd93c5aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.299300 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-kube-api-access-qz6zg" (OuterVolumeSpecName: "kube-api-access-qz6zg") pod "072c5a1a-6292-4420-b5a1-dcd5dd93c5aa" (UID: "072c5a1a-6292-4420-b5a1-dcd5dd93c5aa"). InnerVolumeSpecName "kube-api-access-qz6zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.322687 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-config-data" (OuterVolumeSpecName: "config-data") pod "072c5a1a-6292-4420-b5a1-dcd5dd93c5aa" (UID: "072c5a1a-6292-4420-b5a1-dcd5dd93c5aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.336139 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "072c5a1a-6292-4420-b5a1-dcd5dd93c5aa" (UID: "072c5a1a-6292-4420-b5a1-dcd5dd93c5aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.394585 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.394615 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.394626 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz6zg\" (UniqueName: \"kubernetes.io/projected/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-kube-api-access-qz6zg\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.394634 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.704108 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7bzdn" event={"ID":"072c5a1a-6292-4420-b5a1-dcd5dd93c5aa","Type":"ContainerDied","Data":"35aa72821012d5af9a696c15c60c8d65d4e5d24c1615a9d39fe2ce37d5a88e65"} Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.704164 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35aa72821012d5af9a696c15c60c8d65d4e5d24c1615a9d39fe2ce37d5a88e65" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.704223 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7bzdn" Dec 17 08:42:12 crc kubenswrapper[4966]: I1217 08:42:12.841574 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0f67424-421d-4dc9-baad-af7ad452b801" path="/var/lib/kubelet/pods/c0f67424-421d-4dc9-baad-af7ad452b801/volumes" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.091769 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.092180 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.753968 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qws2d"] Dec 17 08:42:13 crc kubenswrapper[4966]: E1217 08:42:13.754761 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072c5a1a-6292-4420-b5a1-dcd5dd93c5aa" containerName="nova-manage" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.754777 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="072c5a1a-6292-4420-b5a1-dcd5dd93c5aa" containerName="nova-manage" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.754969 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="072c5a1a-6292-4420-b5a1-dcd5dd93c5aa" containerName="nova-manage" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.755687 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.760655 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.760894 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.778819 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qws2d"] Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.926744 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skshc\" (UniqueName: \"kubernetes.io/projected/787797e9-40fd-4108-8448-65dd7ae41e2e-kube-api-access-skshc\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.926833 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-scripts\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.926971 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:13 crc kubenswrapper[4966]: I1217 08:42:13.927019 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-config-data\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.029600 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.029674 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-config-data\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.030763 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skshc\" (UniqueName: \"kubernetes.io/projected/787797e9-40fd-4108-8448-65dd7ae41e2e-kube-api-access-skshc\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.030942 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-scripts\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.035460 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-scripts\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.036276 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.037477 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-config-data\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.052499 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skshc\" (UniqueName: \"kubernetes.io/projected/787797e9-40fd-4108-8448-65dd7ae41e2e-kube-api-access-skshc\") pod \"nova-cell0-cell-mapping-qws2d\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.087019 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.586127 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.663433 4966 trace.go:236] Trace[448979703]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/redhat-operators-c9l6f" (17-Dec-2025 08:42:12.530) (total time: 2133ms): Dec 17 08:42:14 crc kubenswrapper[4966]: Trace[448979703]: [2.133178365s] [2.133178365s] END Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.685501 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c4f4f4d89-wmvz6"] Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.685721 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" podUID="7fb54742-7249-4b48-a918-1bfab858a763" containerName="dnsmasq-dns" containerID="cri-o://3ddc1883fa26c23d1faf2a962987f8ac9f9ac279fdf3b42dc24064cfccdcf8f5" gracePeriod=10 Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.967122 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.974249 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-log" containerID="cri-o://aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c" gracePeriod=30 Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.975134 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-api" containerID="cri-o://a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee" gracePeriod=30 Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.980413 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:42:14 crc kubenswrapper[4966]: I1217 08:42:14.980646 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b3c2e3f9-8b0b-4fb6-8404-035c11650e41" containerName="nova-scheduler-scheduler" containerID="cri-o://cbfcd389f3c7812c50c33f46a05c3b30185207d043879432c77df1ab09c47d21" gracePeriod=30 Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.007824 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.008519 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerName="nova-metadata-log" containerID="cri-o://cb463f8fbc0052ea2ac3e54a53fbd74c5772b2c07972319fd411fcba2eb701f2" gracePeriod=30 Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.008933 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerName="nova-metadata-metadata" containerID="cri-o://97b8c3fecbf1e56e636448b1b4043640278344250085acc71a316af6ef9e2e3d" gracePeriod=30 Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.316434 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:15 crc kubenswrapper[4966]: W1217 08:42:15.337599 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4f6238b_6948_43ae_8a0e_1279ec490a71.slice/crio-077aaaf8f2c5805c1ece5b9bd35d8a817ef02768838c39f6e78dab4442634eaf WatchSource:0}: Error finding container 077aaaf8f2c5805c1ece5b9bd35d8a817ef02768838c39f6e78dab4442634eaf: Status 404 returned error can't find the container with id 077aaaf8f2c5805c1ece5b9bd35d8a817ef02768838c39f6e78dab4442634eaf Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.552206 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qws2d"] Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.829480 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qws2d" event={"ID":"787797e9-40fd-4108-8448-65dd7ae41e2e","Type":"ContainerStarted","Data":"a39f13eee0db2c4171aa432a961c672611af4264122358ae11c40c4775c8aca2"} Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.839151 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerStarted","Data":"077aaaf8f2c5805c1ece5b9bd35d8a817ef02768838c39f6e78dab4442634eaf"} Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.854567 4966 generic.go:334] "Generic (PLEG): container finished" podID="7fb54742-7249-4b48-a918-1bfab858a763" containerID="3ddc1883fa26c23d1faf2a962987f8ac9f9ac279fdf3b42dc24064cfccdcf8f5" exitCode=0 Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.854650 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" event={"ID":"7fb54742-7249-4b48-a918-1bfab858a763","Type":"ContainerDied","Data":"3ddc1883fa26c23d1faf2a962987f8ac9f9ac279fdf3b42dc24064cfccdcf8f5"} Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.858689 4966 generic.go:334] "Generic (PLEG): container finished" podID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerID="aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c" exitCode=143 Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.858739 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8","Type":"ContainerDied","Data":"aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c"} Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.878477 4966 generic.go:334] "Generic (PLEG): container finished" podID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerID="97b8c3fecbf1e56e636448b1b4043640278344250085acc71a316af6ef9e2e3d" exitCode=0 Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.878516 4966 generic.go:334] "Generic (PLEG): container finished" podID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerID="cb463f8fbc0052ea2ac3e54a53fbd74c5772b2c07972319fd411fcba2eb701f2" exitCode=143 Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.878540 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd","Type":"ContainerDied","Data":"97b8c3fecbf1e56e636448b1b4043640278344250085acc71a316af6ef9e2e3d"} Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.878735 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd","Type":"ContainerDied","Data":"cb463f8fbc0052ea2ac3e54a53fbd74c5772b2c07972319fd411fcba2eb701f2"} Dec 17 08:42:15 crc kubenswrapper[4966]: I1217 08:42:15.963257 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.102389 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqzcb\" (UniqueName: \"kubernetes.io/projected/7fb54742-7249-4b48-a918-1bfab858a763-kube-api-access-xqzcb\") pod \"7fb54742-7249-4b48-a918-1bfab858a763\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.102553 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-svc\") pod \"7fb54742-7249-4b48-a918-1bfab858a763\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.102603 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-sb\") pod \"7fb54742-7249-4b48-a918-1bfab858a763\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.102667 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-swift-storage-0\") pod \"7fb54742-7249-4b48-a918-1bfab858a763\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.102762 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-config\") pod \"7fb54742-7249-4b48-a918-1bfab858a763\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.102793 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-nb\") pod \"7fb54742-7249-4b48-a918-1bfab858a763\" (UID: \"7fb54742-7249-4b48-a918-1bfab858a763\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.251317 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fb54742-7249-4b48-a918-1bfab858a763-kube-api-access-xqzcb" (OuterVolumeSpecName: "kube-api-access-xqzcb") pod "7fb54742-7249-4b48-a918-1bfab858a763" (UID: "7fb54742-7249-4b48-a918-1bfab858a763"). InnerVolumeSpecName "kube-api-access-xqzcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.315373 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqzcb\" (UniqueName: \"kubernetes.io/projected/7fb54742-7249-4b48-a918-1bfab858a763-kube-api-access-xqzcb\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.407895 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7fb54742-7249-4b48-a918-1bfab858a763" (UID: "7fb54742-7249-4b48-a918-1bfab858a763"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.422222 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.440853 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7fb54742-7249-4b48-a918-1bfab858a763" (UID: "7fb54742-7249-4b48-a918-1bfab858a763"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.486332 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7fb54742-7249-4b48-a918-1bfab858a763" (UID: "7fb54742-7249-4b48-a918-1bfab858a763"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.489350 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-config" (OuterVolumeSpecName: "config") pod "7fb54742-7249-4b48-a918-1bfab858a763" (UID: "7fb54742-7249-4b48-a918-1bfab858a763"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.519392 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7fb54742-7249-4b48-a918-1bfab858a763" (UID: "7fb54742-7249-4b48-a918-1bfab858a763"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.524035 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.524067 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.524078 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.524088 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb54742-7249-4b48-a918-1bfab858a763-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.531989 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.625038 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw8ck\" (UniqueName: \"kubernetes.io/projected/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-kube-api-access-vw8ck\") pod \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.625073 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-combined-ca-bundle\") pod \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.625170 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-config-data\") pod \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.625232 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-nova-metadata-tls-certs\") pod \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.625277 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-logs\") pod \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\" (UID: \"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd\") " Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.626075 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-logs" (OuterVolumeSpecName: "logs") pod "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" (UID: "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.639021 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-kube-api-access-vw8ck" (OuterVolumeSpecName: "kube-api-access-vw8ck") pod "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" (UID: "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd"). InnerVolumeSpecName "kube-api-access-vw8ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.675752 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-config-data" (OuterVolumeSpecName: "config-data") pod "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" (UID: "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.714650 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" (UID: "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.727092 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" (UID: "3b99d0d7-30a1-4d05-996a-38a64e1e2bfd"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.728341 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw8ck\" (UniqueName: \"kubernetes.io/projected/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-kube-api-access-vw8ck\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.728365 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.728376 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.728387 4966 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.728395 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.812998 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.813350 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.813393 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.814316 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ede17bc7a2bbfa2ac454efc1ee7643bc3db24c460d6b848c803354fff3d0b2f2"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.814395 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://ede17bc7a2bbfa2ac454efc1ee7643bc3db24c460d6b848c803354fff3d0b2f2" gracePeriod=600 Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.922332 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" event={"ID":"7fb54742-7249-4b48-a918-1bfab858a763","Type":"ContainerDied","Data":"077204e63e9658c00f670a0e802a7378f50202342a418d8a943cffcd1382b609"} Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.922382 4966 scope.go:117] "RemoveContainer" containerID="3ddc1883fa26c23d1faf2a962987f8ac9f9ac279fdf3b42dc24064cfccdcf8f5" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.922503 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c4f4f4d89-wmvz6" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.926847 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b99d0d7-30a1-4d05-996a-38a64e1e2bfd","Type":"ContainerDied","Data":"6da804fdcee0f8c1c554b946537375c761e2482ec16cad6838f5bea4cac570c5"} Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.926957 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.944446 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qws2d" event={"ID":"787797e9-40fd-4108-8448-65dd7ae41e2e","Type":"ContainerStarted","Data":"069702d674a0dfc7ece303cf410082fe66470f695b0900569072234a90df79f9"} Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.964018 4966 scope.go:117] "RemoveContainer" containerID="a4a4f9b441601fb5b2ec86eb01abdb18b1fb6827411014b5de9031bbe0515bd9" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.964576 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c4f4f4d89-wmvz6"] Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.964639 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerStarted","Data":"b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e"} Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.964663 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerStarted","Data":"cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f"} Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.983381 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qws2d" podStartSLOduration=3.983365302 podStartE2EDuration="3.983365302s" podCreationTimestamp="2025-12-17 08:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:16.967820595 +0000 UTC m=+1272.512890537" watchObservedRunningTime="2025-12-17 08:42:16.983365302 +0000 UTC m=+1272.528435244" Dec 17 08:42:16 crc kubenswrapper[4966]: I1217 08:42:16.986948 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c4f4f4d89-wmvz6"] Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.048970 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.132095 4966 scope.go:117] "RemoveContainer" containerID="97b8c3fecbf1e56e636448b1b4043640278344250085acc71a316af6ef9e2e3d" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.144949 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.191566 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:17 crc kubenswrapper[4966]: E1217 08:42:17.192150 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerName="nova-metadata-log" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.192170 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerName="nova-metadata-log" Dec 17 08:42:17 crc kubenswrapper[4966]: E1217 08:42:17.192188 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb54742-7249-4b48-a918-1bfab858a763" containerName="init" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.192194 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb54742-7249-4b48-a918-1bfab858a763" containerName="init" Dec 17 08:42:17 crc kubenswrapper[4966]: E1217 08:42:17.192209 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb54742-7249-4b48-a918-1bfab858a763" containerName="dnsmasq-dns" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.192215 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb54742-7249-4b48-a918-1bfab858a763" containerName="dnsmasq-dns" Dec 17 08:42:17 crc kubenswrapper[4966]: E1217 08:42:17.192229 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerName="nova-metadata-metadata" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.192235 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerName="nova-metadata-metadata" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.192451 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerName="nova-metadata-log" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.192475 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fb54742-7249-4b48-a918-1bfab858a763" containerName="dnsmasq-dns" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.192487 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" containerName="nova-metadata-metadata" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.193588 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.198394 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.199210 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.231331 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.288240 4966 scope.go:117] "RemoveContainer" containerID="cb463f8fbc0052ea2ac3e54a53fbd74c5772b2c07972319fd411fcba2eb701f2" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.360645 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-logs\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.360697 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.360840 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-config-data\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.361584 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjbwx\" (UniqueName: \"kubernetes.io/projected/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-kube-api-access-sjbwx\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.361621 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.463250 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-config-data\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.463375 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjbwx\" (UniqueName: \"kubernetes.io/projected/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-kube-api-access-sjbwx\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.463429 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.463510 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-logs\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.463529 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.464726 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-logs\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.472569 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.473407 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.474832 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-config-data\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.495565 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjbwx\" (UniqueName: \"kubernetes.io/projected/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-kube-api-access-sjbwx\") pod \"nova-metadata-0\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.721153 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.991589 4966 generic.go:334] "Generic (PLEG): container finished" podID="b3c2e3f9-8b0b-4fb6-8404-035c11650e41" containerID="cbfcd389f3c7812c50c33f46a05c3b30185207d043879432c77df1ab09c47d21" exitCode=0 Dec 17 08:42:17 crc kubenswrapper[4966]: I1217 08:42:17.992320 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3c2e3f9-8b0b-4fb6-8404-035c11650e41","Type":"ContainerDied","Data":"cbfcd389f3c7812c50c33f46a05c3b30185207d043879432c77df1ab09c47d21"} Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.024781 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerStarted","Data":"44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297"} Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.051201 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="ede17bc7a2bbfa2ac454efc1ee7643bc3db24c460d6b848c803354fff3d0b2f2" exitCode=0 Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.051284 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"ede17bc7a2bbfa2ac454efc1ee7643bc3db24c460d6b848c803354fff3d0b2f2"} Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.051329 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"682169a45fcdda5953065f4e8d118ba911e57a4b24a9e7bfd550631091e0d37e"} Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.051347 4966 scope.go:117] "RemoveContainer" containerID="dfe84a49239318b4c24aec015fc04c7e7ced77b13085152134e648f1545c4edf" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.297532 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:18 crc kubenswrapper[4966]: W1217 08:42:18.363004 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3b5f2dd_052a_4bb8_afaf_96b9ef596219.slice/crio-4ea51ce368438542045aea8f1dd61eff8fd9f4bfc65b5d2380d87ee3ddb49ee6 WatchSource:0}: Error finding container 4ea51ce368438542045aea8f1dd61eff8fd9f4bfc65b5d2380d87ee3ddb49ee6: Status 404 returned error can't find the container with id 4ea51ce368438542045aea8f1dd61eff8fd9f4bfc65b5d2380d87ee3ddb49ee6 Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.519323 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.589892 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5qxw\" (UniqueName: \"kubernetes.io/projected/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-kube-api-access-q5qxw\") pod \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.590124 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-combined-ca-bundle\") pod \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.590159 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-config-data\") pod \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\" (UID: \"b3c2e3f9-8b0b-4fb6-8404-035c11650e41\") " Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.598335 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-kube-api-access-q5qxw" (OuterVolumeSpecName: "kube-api-access-q5qxw") pod "b3c2e3f9-8b0b-4fb6-8404-035c11650e41" (UID: "b3c2e3f9-8b0b-4fb6-8404-035c11650e41"). InnerVolumeSpecName "kube-api-access-q5qxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.687706 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3c2e3f9-8b0b-4fb6-8404-035c11650e41" (UID: "b3c2e3f9-8b0b-4fb6-8404-035c11650e41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.701120 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5qxw\" (UniqueName: \"kubernetes.io/projected/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-kube-api-access-q5qxw\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.701153 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.851268 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b99d0d7-30a1-4d05-996a-38a64e1e2bfd" path="/var/lib/kubelet/pods/3b99d0d7-30a1-4d05-996a-38a64e1e2bfd/volumes" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.852252 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fb54742-7249-4b48-a918-1bfab858a763" path="/var/lib/kubelet/pods/7fb54742-7249-4b48-a918-1bfab858a763/volumes" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.860481 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-config-data" (OuterVolumeSpecName: "config-data") pod "b3c2e3f9-8b0b-4fb6-8404-035c11650e41" (UID: "b3c2e3f9-8b0b-4fb6-8404-035c11650e41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.863249 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:42:18 crc kubenswrapper[4966]: I1217 08:42:18.906285 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3c2e3f9-8b0b-4fb6-8404-035c11650e41-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.007262 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj458\" (UniqueName: \"kubernetes.io/projected/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-kube-api-access-pj458\") pod \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.007411 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-logs\") pod \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.007512 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-config-data\") pod \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.008243 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-logs" (OuterVolumeSpecName: "logs") pod "248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" (UID: "248f6a8e-1c9c-42e7-9e16-9f16aa6049f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.008324 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-combined-ca-bundle\") pod \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\" (UID: \"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8\") " Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.008780 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.012827 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-kube-api-access-pj458" (OuterVolumeSpecName: "kube-api-access-pj458") pod "248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" (UID: "248f6a8e-1c9c-42e7-9e16-9f16aa6049f8"). InnerVolumeSpecName "kube-api-access-pj458". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.052197 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" (UID: "248f6a8e-1c9c-42e7-9e16-9f16aa6049f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.054966 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-config-data" (OuterVolumeSpecName: "config-data") pod "248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" (UID: "248f6a8e-1c9c-42e7-9e16-9f16aa6049f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.074463 4966 generic.go:334] "Generic (PLEG): container finished" podID="a272c7ab-dc3f-49af-9f02-6fe4b552c217" containerID="4eb94a4fb5c18d6acd037c4294a3352d575e3b79ac2e5f740d46fa74aa105530" exitCode=0 Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.074748 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tdhl5" event={"ID":"a272c7ab-dc3f-49af-9f02-6fe4b552c217","Type":"ContainerDied","Data":"4eb94a4fb5c18d6acd037c4294a3352d575e3b79ac2e5f740d46fa74aa105530"} Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.087211 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d3b5f2dd-052a-4bb8-afaf-96b9ef596219","Type":"ContainerStarted","Data":"035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32"} Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.087471 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d3b5f2dd-052a-4bb8-afaf-96b9ef596219","Type":"ContainerStarted","Data":"4ea51ce368438542045aea8f1dd61eff8fd9f4bfc65b5d2380d87ee3ddb49ee6"} Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.103335 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b3c2e3f9-8b0b-4fb6-8404-035c11650e41","Type":"ContainerDied","Data":"2e52d97b0e92ae614d8dead8675ca6e39d275cea344c8c789ab851b697238063"} Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.103379 4966 scope.go:117] "RemoveContainer" containerID="cbfcd389f3c7812c50c33f46a05c3b30185207d043879432c77df1ab09c47d21" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.103469 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.113212 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj458\" (UniqueName: \"kubernetes.io/projected/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-kube-api-access-pj458\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.113331 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.113409 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.128285 4966 generic.go:334] "Generic (PLEG): container finished" podID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerID="a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee" exitCode=0 Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.128513 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8","Type":"ContainerDied","Data":"a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee"} Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.128604 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"248f6a8e-1c9c-42e7-9e16-9f16aa6049f8","Type":"ContainerDied","Data":"d757f6a7640c90e29fa36db8c5b81ae51adb9c1adfe54b0ea0b5cdeee9ece6cc"} Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.128743 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.203821 4966 scope.go:117] "RemoveContainer" containerID="a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.232773 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.256005 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.267377 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:19 crc kubenswrapper[4966]: E1217 08:42:19.267766 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c2e3f9-8b0b-4fb6-8404-035c11650e41" containerName="nova-scheduler-scheduler" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.267780 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c2e3f9-8b0b-4fb6-8404-035c11650e41" containerName="nova-scheduler-scheduler" Dec 17 08:42:19 crc kubenswrapper[4966]: E1217 08:42:19.267793 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-log" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.267801 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-log" Dec 17 08:42:19 crc kubenswrapper[4966]: E1217 08:42:19.267842 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-api" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.267849 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-api" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.268030 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-log" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.268061 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" containerName="nova-api-api" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.268073 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c2e3f9-8b0b-4fb6-8404-035c11650e41" containerName="nova-scheduler-scheduler" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.269049 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.274448 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.290517 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.291032 4966 scope.go:117] "RemoveContainer" containerID="aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.322947 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.333105 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.360924 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.362279 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.367267 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.390863 4966 scope.go:117] "RemoveContainer" containerID="a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee" Dec 17 08:42:19 crc kubenswrapper[4966]: E1217 08:42:19.392641 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee\": container with ID starting with a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee not found: ID does not exist" containerID="a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.392684 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee"} err="failed to get container status \"a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee\": rpc error: code = NotFound desc = could not find container \"a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee\": container with ID starting with a7fa44cf877af09ba683fa84faf13a088220069b4b6a7016b4d38e4ab175e5ee not found: ID does not exist" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.392712 4966 scope.go:117] "RemoveContainer" containerID="aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c" Dec 17 08:42:19 crc kubenswrapper[4966]: E1217 08:42:19.394372 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c\": container with ID starting with aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c not found: ID does not exist" containerID="aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.394395 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c"} err="failed to get container status \"aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c\": rpc error: code = NotFound desc = could not find container \"aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c\": container with ID starting with aeab64aba63637f58b5c07f953974b47ed068f713a179b13e7724fedf72e782c not found: ID does not exist" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.409810 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.435180 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.435271 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9014cabb-e109-4aed-99c8-990896bd10bf-logs\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.435305 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfzt8\" (UniqueName: \"kubernetes.io/projected/9014cabb-e109-4aed-99c8-990896bd10bf-kube-api-access-nfzt8\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.435442 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-config-data\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.536852 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-config-data\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.536933 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.537011 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9014cabb-e109-4aed-99c8-990896bd10bf-logs\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.537035 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfzt8\" (UniqueName: \"kubernetes.io/projected/9014cabb-e109-4aed-99c8-990896bd10bf-kube-api-access-nfzt8\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.537087 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.537142 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trjpt\" (UniqueName: \"kubernetes.io/projected/03b5489a-56c8-441a-8fb6-421ac6275281-kube-api-access-trjpt\") pod \"nova-scheduler-0\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.537165 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-config-data\") pod \"nova-scheduler-0\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.538558 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9014cabb-e109-4aed-99c8-990896bd10bf-logs\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.550823 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.565576 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-config-data\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.567406 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfzt8\" (UniqueName: \"kubernetes.io/projected/9014cabb-e109-4aed-99c8-990896bd10bf-kube-api-access-nfzt8\") pod \"nova-api-0\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.584696 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.642741 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trjpt\" (UniqueName: \"kubernetes.io/projected/03b5489a-56c8-441a-8fb6-421ac6275281-kube-api-access-trjpt\") pod \"nova-scheduler-0\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.642793 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-config-data\") pod \"nova-scheduler-0\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.642950 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.654569 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-config-data\") pod \"nova-scheduler-0\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.655281 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.668624 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trjpt\" (UniqueName: \"kubernetes.io/projected/03b5489a-56c8-441a-8fb6-421ac6275281-kube-api-access-trjpt\") pod \"nova-scheduler-0\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " pod="openstack/nova-scheduler-0" Dec 17 08:42:19 crc kubenswrapper[4966]: I1217 08:42:19.705858 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.134452 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.142585 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerStarted","Data":"473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780"} Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.142715 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 17 08:42:20 crc kubenswrapper[4966]: W1217 08:42:20.143720 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9014cabb_e109_4aed_99c8_990896bd10bf.slice/crio-5ba6d85420f5904d4c2cc531c6e2bcfccbf55a1482a95165f89303be8d405cce WatchSource:0}: Error finding container 5ba6d85420f5904d4c2cc531c6e2bcfccbf55a1482a95165f89303be8d405cce: Status 404 returned error can't find the container with id 5ba6d85420f5904d4c2cc531c6e2bcfccbf55a1482a95165f89303be8d405cce Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.149023 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d3b5f2dd-052a-4bb8-afaf-96b9ef596219","Type":"ContainerStarted","Data":"048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e"} Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.177518 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.520835611 podStartE2EDuration="9.177494572s" podCreationTimestamp="2025-12-17 08:42:11 +0000 UTC" firstStartedPulling="2025-12-17 08:42:15.34336602 +0000 UTC m=+1270.888435962" lastFinishedPulling="2025-12-17 08:42:19.000024971 +0000 UTC m=+1274.545094923" observedRunningTime="2025-12-17 08:42:20.169329918 +0000 UTC m=+1275.714399870" watchObservedRunningTime="2025-12-17 08:42:20.177494572 +0000 UTC m=+1275.722564514" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.201288 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.201253984 podStartE2EDuration="4.201253984s" podCreationTimestamp="2025-12-17 08:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:20.196518843 +0000 UTC m=+1275.741588775" watchObservedRunningTime="2025-12-17 08:42:20.201253984 +0000 UTC m=+1275.746323926" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.257910 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.549987 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.681731 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-scripts\") pod \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.681862 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rmtv\" (UniqueName: \"kubernetes.io/projected/a272c7ab-dc3f-49af-9f02-6fe4b552c217-kube-api-access-2rmtv\") pod \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.681931 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-combined-ca-bundle\") pod \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.681963 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-config-data\") pod \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\" (UID: \"a272c7ab-dc3f-49af-9f02-6fe4b552c217\") " Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.696006 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-scripts" (OuterVolumeSpecName: "scripts") pod "a272c7ab-dc3f-49af-9f02-6fe4b552c217" (UID: "a272c7ab-dc3f-49af-9f02-6fe4b552c217"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.696119 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a272c7ab-dc3f-49af-9f02-6fe4b552c217-kube-api-access-2rmtv" (OuterVolumeSpecName: "kube-api-access-2rmtv") pod "a272c7ab-dc3f-49af-9f02-6fe4b552c217" (UID: "a272c7ab-dc3f-49af-9f02-6fe4b552c217"). InnerVolumeSpecName "kube-api-access-2rmtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.719270 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a272c7ab-dc3f-49af-9f02-6fe4b552c217" (UID: "a272c7ab-dc3f-49af-9f02-6fe4b552c217"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.722349 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-config-data" (OuterVolumeSpecName: "config-data") pod "a272c7ab-dc3f-49af-9f02-6fe4b552c217" (UID: "a272c7ab-dc3f-49af-9f02-6fe4b552c217"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.783645 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.783675 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rmtv\" (UniqueName: \"kubernetes.io/projected/a272c7ab-dc3f-49af-9f02-6fe4b552c217-kube-api-access-2rmtv\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.783687 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.783695 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a272c7ab-dc3f-49af-9f02-6fe4b552c217-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.842329 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="248f6a8e-1c9c-42e7-9e16-9f16aa6049f8" path="/var/lib/kubelet/pods/248f6a8e-1c9c-42e7-9e16-9f16aa6049f8/volumes" Dec 17 08:42:20 crc kubenswrapper[4966]: I1217 08:42:20.843285 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3c2e3f9-8b0b-4fb6-8404-035c11650e41" path="/var/lib/kubelet/pods/b3c2e3f9-8b0b-4fb6-8404-035c11650e41/volumes" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.208664 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 17 08:42:21 crc kubenswrapper[4966]: E1217 08:42:21.209311 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a272c7ab-dc3f-49af-9f02-6fe4b552c217" containerName="nova-cell1-conductor-db-sync" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.209323 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a272c7ab-dc3f-49af-9f02-6fe4b552c217" containerName="nova-cell1-conductor-db-sync" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.209534 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a272c7ab-dc3f-49af-9f02-6fe4b552c217" containerName="nova-cell1-conductor-db-sync" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.215928 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.224699 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.254971 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9014cabb-e109-4aed-99c8-990896bd10bf","Type":"ContainerStarted","Data":"0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f"} Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.255009 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9014cabb-e109-4aed-99c8-990896bd10bf","Type":"ContainerStarted","Data":"9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a"} Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.255021 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9014cabb-e109-4aed-99c8-990896bd10bf","Type":"ContainerStarted","Data":"5ba6d85420f5904d4c2cc531c6e2bcfccbf55a1482a95165f89303be8d405cce"} Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.262774 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"03b5489a-56c8-441a-8fb6-421ac6275281","Type":"ContainerStarted","Data":"acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e"} Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.262824 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"03b5489a-56c8-441a-8fb6-421ac6275281","Type":"ContainerStarted","Data":"25e03a3af423b14658801279e13115dec81085d56fa0fd2e3b4fd1f91a6ccd22"} Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.268378 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tdhl5" event={"ID":"a272c7ab-dc3f-49af-9f02-6fe4b552c217","Type":"ContainerDied","Data":"17dee0eec6af7b7cb23582eac74d110f00449db17cbdce27fb78f84bd2c0f8d9"} Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.268428 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17dee0eec6af7b7cb23582eac74d110f00449db17cbdce27fb78f84bd2c0f8d9" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.269092 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tdhl5" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.299691 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.299670634 podStartE2EDuration="2.299670634s" podCreationTimestamp="2025-12-17 08:42:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:21.279271654 +0000 UTC m=+1276.824341596" watchObservedRunningTime="2025-12-17 08:42:21.299670634 +0000 UTC m=+1276.844740576" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.314333 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.314310467 podStartE2EDuration="2.314310467s" podCreationTimestamp="2025-12-17 08:42:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:21.293156555 +0000 UTC m=+1276.838226497" watchObservedRunningTime="2025-12-17 08:42:21.314310467 +0000 UTC m=+1276.859380429" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.395988 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe406a2-f0ae-40c0-9bf2-dd20761ceed2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"efe406a2-f0ae-40c0-9bf2-dd20761ceed2\") " pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.396100 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w4tr\" (UniqueName: \"kubernetes.io/projected/efe406a2-f0ae-40c0-9bf2-dd20761ceed2-kube-api-access-9w4tr\") pod \"nova-cell1-conductor-0\" (UID: \"efe406a2-f0ae-40c0-9bf2-dd20761ceed2\") " pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.396947 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe406a2-f0ae-40c0-9bf2-dd20761ceed2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"efe406a2-f0ae-40c0-9bf2-dd20761ceed2\") " pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.498420 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe406a2-f0ae-40c0-9bf2-dd20761ceed2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"efe406a2-f0ae-40c0-9bf2-dd20761ceed2\") " pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.498546 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe406a2-f0ae-40c0-9bf2-dd20761ceed2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"efe406a2-f0ae-40c0-9bf2-dd20761ceed2\") " pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.499220 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w4tr\" (UniqueName: \"kubernetes.io/projected/efe406a2-f0ae-40c0-9bf2-dd20761ceed2-kube-api-access-9w4tr\") pod \"nova-cell1-conductor-0\" (UID: \"efe406a2-f0ae-40c0-9bf2-dd20761ceed2\") " pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.502550 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe406a2-f0ae-40c0-9bf2-dd20761ceed2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"efe406a2-f0ae-40c0-9bf2-dd20761ceed2\") " pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.515533 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe406a2-f0ae-40c0-9bf2-dd20761ceed2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"efe406a2-f0ae-40c0-9bf2-dd20761ceed2\") " pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.518411 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w4tr\" (UniqueName: \"kubernetes.io/projected/efe406a2-f0ae-40c0-9bf2-dd20761ceed2-kube-api-access-9w4tr\") pod \"nova-cell1-conductor-0\" (UID: \"efe406a2-f0ae-40c0-9bf2-dd20761ceed2\") " pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:21 crc kubenswrapper[4966]: I1217 08:42:21.565071 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:22 crc kubenswrapper[4966]: I1217 08:42:22.151253 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 17 08:42:22 crc kubenswrapper[4966]: I1217 08:42:22.302210 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"efe406a2-f0ae-40c0-9bf2-dd20761ceed2","Type":"ContainerStarted","Data":"70e3af8d044944e2f44d0a30a05862c52ad64c28d892289ea74ad37619d0cea0"} Dec 17 08:42:22 crc kubenswrapper[4966]: I1217 08:42:22.722182 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 17 08:42:22 crc kubenswrapper[4966]: I1217 08:42:22.722548 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 17 08:42:23 crc kubenswrapper[4966]: I1217 08:42:23.316913 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"efe406a2-f0ae-40c0-9bf2-dd20761ceed2","Type":"ContainerStarted","Data":"f0285f6118f175659456e197855403fc77512840df29b3fd548308ad7365b88d"} Dec 17 08:42:23 crc kubenswrapper[4966]: I1217 08:42:23.318193 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:23 crc kubenswrapper[4966]: I1217 08:42:23.349127 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.349107542 podStartE2EDuration="2.349107542s" podCreationTimestamp="2025-12-17 08:42:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:23.331493209 +0000 UTC m=+1278.876563161" watchObservedRunningTime="2025-12-17 08:42:23.349107542 +0000 UTC m=+1278.894177504" Dec 17 08:42:24 crc kubenswrapper[4966]: I1217 08:42:24.328889 4966 generic.go:334] "Generic (PLEG): container finished" podID="787797e9-40fd-4108-8448-65dd7ae41e2e" containerID="069702d674a0dfc7ece303cf410082fe66470f695b0900569072234a90df79f9" exitCode=0 Dec 17 08:42:24 crc kubenswrapper[4966]: I1217 08:42:24.328997 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qws2d" event={"ID":"787797e9-40fd-4108-8448-65dd7ae41e2e","Type":"ContainerDied","Data":"069702d674a0dfc7ece303cf410082fe66470f695b0900569072234a90df79f9"} Dec 17 08:42:24 crc kubenswrapper[4966]: I1217 08:42:24.707238 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 17 08:42:25 crc kubenswrapper[4966]: I1217 08:42:25.728108 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:25 crc kubenswrapper[4966]: I1217 08:42:25.905263 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-config-data\") pod \"787797e9-40fd-4108-8448-65dd7ae41e2e\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " Dec 17 08:42:25 crc kubenswrapper[4966]: I1217 08:42:25.905506 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-combined-ca-bundle\") pod \"787797e9-40fd-4108-8448-65dd7ae41e2e\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " Dec 17 08:42:25 crc kubenswrapper[4966]: I1217 08:42:25.905562 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skshc\" (UniqueName: \"kubernetes.io/projected/787797e9-40fd-4108-8448-65dd7ae41e2e-kube-api-access-skshc\") pod \"787797e9-40fd-4108-8448-65dd7ae41e2e\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " Dec 17 08:42:25 crc kubenswrapper[4966]: I1217 08:42:25.905626 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-scripts\") pod \"787797e9-40fd-4108-8448-65dd7ae41e2e\" (UID: \"787797e9-40fd-4108-8448-65dd7ae41e2e\") " Dec 17 08:42:25 crc kubenswrapper[4966]: I1217 08:42:25.911933 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/787797e9-40fd-4108-8448-65dd7ae41e2e-kube-api-access-skshc" (OuterVolumeSpecName: "kube-api-access-skshc") pod "787797e9-40fd-4108-8448-65dd7ae41e2e" (UID: "787797e9-40fd-4108-8448-65dd7ae41e2e"). InnerVolumeSpecName "kube-api-access-skshc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:25 crc kubenswrapper[4966]: I1217 08:42:25.914025 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-scripts" (OuterVolumeSpecName: "scripts") pod "787797e9-40fd-4108-8448-65dd7ae41e2e" (UID: "787797e9-40fd-4108-8448-65dd7ae41e2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:25 crc kubenswrapper[4966]: I1217 08:42:25.939065 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-config-data" (OuterVolumeSpecName: "config-data") pod "787797e9-40fd-4108-8448-65dd7ae41e2e" (UID: "787797e9-40fd-4108-8448-65dd7ae41e2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:25 crc kubenswrapper[4966]: I1217 08:42:25.942512 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "787797e9-40fd-4108-8448-65dd7ae41e2e" (UID: "787797e9-40fd-4108-8448-65dd7ae41e2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:26 crc kubenswrapper[4966]: I1217 08:42:26.007812 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:26 crc kubenswrapper[4966]: I1217 08:42:26.007845 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:26 crc kubenswrapper[4966]: I1217 08:42:26.007858 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skshc\" (UniqueName: \"kubernetes.io/projected/787797e9-40fd-4108-8448-65dd7ae41e2e-kube-api-access-skshc\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:26 crc kubenswrapper[4966]: I1217 08:42:26.007936 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/787797e9-40fd-4108-8448-65dd7ae41e2e-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:26 crc kubenswrapper[4966]: I1217 08:42:26.350383 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qws2d" event={"ID":"787797e9-40fd-4108-8448-65dd7ae41e2e","Type":"ContainerDied","Data":"a39f13eee0db2c4171aa432a961c672611af4264122358ae11c40c4775c8aca2"} Dec 17 08:42:26 crc kubenswrapper[4966]: I1217 08:42:26.350416 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a39f13eee0db2c4171aa432a961c672611af4264122358ae11c40c4775c8aca2" Dec 17 08:42:26 crc kubenswrapper[4966]: I1217 08:42:26.350479 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qws2d" Dec 17 08:42:27 crc kubenswrapper[4966]: I1217 08:42:27.721690 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 17 08:42:27 crc kubenswrapper[4966]: I1217 08:42:27.722191 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 17 08:42:28 crc kubenswrapper[4966]: I1217 08:42:28.746084 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 08:42:28 crc kubenswrapper[4966]: I1217 08:42:28.746659 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 08:42:29 crc kubenswrapper[4966]: I1217 08:42:29.585274 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 17 08:42:29 crc kubenswrapper[4966]: I1217 08:42:29.585316 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 17 08:42:29 crc kubenswrapper[4966]: I1217 08:42:29.706938 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 17 08:42:29 crc kubenswrapper[4966]: I1217 08:42:29.754370 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 17 08:42:30 crc kubenswrapper[4966]: I1217 08:42:30.459562 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 17 08:42:30 crc kubenswrapper[4966]: I1217 08:42:30.668176 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.212:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 08:42:30 crc kubenswrapper[4966]: I1217 08:42:30.668727 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.212:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 08:42:31 crc kubenswrapper[4966]: I1217 08:42:31.594635 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 17 08:42:35 crc kubenswrapper[4966]: I1217 08:42:35.977238 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.137903 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-config-data\") pod \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.138354 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htn8h\" (UniqueName: \"kubernetes.io/projected/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-kube-api-access-htn8h\") pod \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.138496 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-combined-ca-bundle\") pod \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\" (UID: \"4a5ceb27-fa91-498b-90d6-1828d9efbf5f\") " Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.146836 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-kube-api-access-htn8h" (OuterVolumeSpecName: "kube-api-access-htn8h") pod "4a5ceb27-fa91-498b-90d6-1828d9efbf5f" (UID: "4a5ceb27-fa91-498b-90d6-1828d9efbf5f"). InnerVolumeSpecName "kube-api-access-htn8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.172505 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a5ceb27-fa91-498b-90d6-1828d9efbf5f" (UID: "4a5ceb27-fa91-498b-90d6-1828d9efbf5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.187554 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-config-data" (OuterVolumeSpecName: "config-data") pod "4a5ceb27-fa91-498b-90d6-1828d9efbf5f" (UID: "4a5ceb27-fa91-498b-90d6-1828d9efbf5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.240423 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.240461 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htn8h\" (UniqueName: \"kubernetes.io/projected/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-kube-api-access-htn8h\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.240473 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5ceb27-fa91-498b-90d6-1828d9efbf5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.457593 4966 generic.go:334] "Generic (PLEG): container finished" podID="4a5ceb27-fa91-498b-90d6-1828d9efbf5f" containerID="60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a" exitCode=137 Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.457638 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4a5ceb27-fa91-498b-90d6-1828d9efbf5f","Type":"ContainerDied","Data":"60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a"} Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.457668 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4a5ceb27-fa91-498b-90d6-1828d9efbf5f","Type":"ContainerDied","Data":"d1459b38e7df3398f23791cad7435efa32e68fe980aa50619f343b5c68d7bc00"} Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.457690 4966 scope.go:117] "RemoveContainer" containerID="60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.457811 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.495023 4966 scope.go:117] "RemoveContainer" containerID="60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a" Dec 17 08:42:36 crc kubenswrapper[4966]: E1217 08:42:36.495727 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a\": container with ID starting with 60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a not found: ID does not exist" containerID="60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.495758 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a"} err="failed to get container status \"60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a\": rpc error: code = NotFound desc = could not find container \"60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a\": container with ID starting with 60db4eb52e6c33cbd73ddf76187bec6c731e0eee10d7088f05536c4390eb695a not found: ID does not exist" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.496922 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.511375 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.526618 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 17 08:42:36 crc kubenswrapper[4966]: E1217 08:42:36.533540 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="787797e9-40fd-4108-8448-65dd7ae41e2e" containerName="nova-manage" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.533583 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="787797e9-40fd-4108-8448-65dd7ae41e2e" containerName="nova-manage" Dec 17 08:42:36 crc kubenswrapper[4966]: E1217 08:42:36.533611 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a5ceb27-fa91-498b-90d6-1828d9efbf5f" containerName="nova-cell1-novncproxy-novncproxy" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.533619 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a5ceb27-fa91-498b-90d6-1828d9efbf5f" containerName="nova-cell1-novncproxy-novncproxy" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.533956 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a5ceb27-fa91-498b-90d6-1828d9efbf5f" containerName="nova-cell1-novncproxy-novncproxy" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.533981 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="787797e9-40fd-4108-8448-65dd7ae41e2e" containerName="nova-manage" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.537403 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.541173 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.541483 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.541681 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.565650 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.648886 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k8ns\" (UniqueName: \"kubernetes.io/projected/4cc357dd-7994-4056-a964-ee9274385567-kube-api-access-4k8ns\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.649208 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.649450 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.649822 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.649949 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.751992 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.752036 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.752055 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.752089 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k8ns\" (UniqueName: \"kubernetes.io/projected/4cc357dd-7994-4056-a964-ee9274385567-kube-api-access-4k8ns\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.752117 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.755713 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.756130 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.756622 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.762415 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cc357dd-7994-4056-a964-ee9274385567-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.769313 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k8ns\" (UniqueName: \"kubernetes.io/projected/4cc357dd-7994-4056-a964-ee9274385567-kube-api-access-4k8ns\") pod \"nova-cell1-novncproxy-0\" (UID: \"4cc357dd-7994-4056-a964-ee9274385567\") " pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.844559 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a5ceb27-fa91-498b-90d6-1828d9efbf5f" path="/var/lib/kubelet/pods/4a5ceb27-fa91-498b-90d6-1828d9efbf5f/volumes" Dec 17 08:42:36 crc kubenswrapper[4966]: I1217 08:42:36.860315 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:37 crc kubenswrapper[4966]: I1217 08:42:37.413884 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 17 08:42:37 crc kubenswrapper[4966]: W1217 08:42:37.417131 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cc357dd_7994_4056_a964_ee9274385567.slice/crio-3da19c6edb6a7ff47768dea8d57771f333127ba19c208a20e35a944efbc36aa7 WatchSource:0}: Error finding container 3da19c6edb6a7ff47768dea8d57771f333127ba19c208a20e35a944efbc36aa7: Status 404 returned error can't find the container with id 3da19c6edb6a7ff47768dea8d57771f333127ba19c208a20e35a944efbc36aa7 Dec 17 08:42:37 crc kubenswrapper[4966]: I1217 08:42:37.473305 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4cc357dd-7994-4056-a964-ee9274385567","Type":"ContainerStarted","Data":"3da19c6edb6a7ff47768dea8d57771f333127ba19c208a20e35a944efbc36aa7"} Dec 17 08:42:37 crc kubenswrapper[4966]: I1217 08:42:37.736498 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 17 08:42:37 crc kubenswrapper[4966]: I1217 08:42:37.736924 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 17 08:42:37 crc kubenswrapper[4966]: I1217 08:42:37.742205 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 17 08:42:37 crc kubenswrapper[4966]: I1217 08:42:37.743558 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 17 08:42:38 crc kubenswrapper[4966]: I1217 08:42:38.484923 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4cc357dd-7994-4056-a964-ee9274385567","Type":"ContainerStarted","Data":"4b9247cf4dd50c8ca9f64f5f6558d13aafb57da2fb49cdbe63797cca4b3abbf6"} Dec 17 08:42:38 crc kubenswrapper[4966]: I1217 08:42:38.513206 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.513187641 podStartE2EDuration="2.513187641s" podCreationTimestamp="2025-12-17 08:42:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:38.512461922 +0000 UTC m=+1294.057531874" watchObservedRunningTime="2025-12-17 08:42:38.513187641 +0000 UTC m=+1294.058257594" Dec 17 08:42:39 crc kubenswrapper[4966]: I1217 08:42:39.590688 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 17 08:42:39 crc kubenswrapper[4966]: I1217 08:42:39.591048 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 17 08:42:39 crc kubenswrapper[4966]: I1217 08:42:39.591572 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 17 08:42:39 crc kubenswrapper[4966]: I1217 08:42:39.598601 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.502215 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.510339 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.765574 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d65fcb98c-nd4qw"] Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.768001 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.778296 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d65fcb98c-nd4qw"] Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.870575 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-sb\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.870683 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-svc\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.870742 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-config\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.870796 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxv4b\" (UniqueName: \"kubernetes.io/projected/b64018c5-5835-4e79-908f-b2f96d279dcf-kube-api-access-wxv4b\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.870832 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-swift-storage-0\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.870893 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-nb\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.972558 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxv4b\" (UniqueName: \"kubernetes.io/projected/b64018c5-5835-4e79-908f-b2f96d279dcf-kube-api-access-wxv4b\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.972625 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-swift-storage-0\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.972673 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-nb\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.972744 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-sb\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.972818 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-svc\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.972962 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-config\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.974315 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-swift-storage-0\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.974715 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-svc\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.974958 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-nb\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.975734 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-config\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.975817 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-sb\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:40 crc kubenswrapper[4966]: I1217 08:42:40.997559 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxv4b\" (UniqueName: \"kubernetes.io/projected/b64018c5-5835-4e79-908f-b2f96d279dcf-kube-api-access-wxv4b\") pod \"dnsmasq-dns-5d65fcb98c-nd4qw\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:41 crc kubenswrapper[4966]: I1217 08:42:41.087632 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:41 crc kubenswrapper[4966]: I1217 08:42:41.570207 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d65fcb98c-nd4qw"] Dec 17 08:42:41 crc kubenswrapper[4966]: I1217 08:42:41.862410 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:42 crc kubenswrapper[4966]: I1217 08:42:42.181236 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 17 08:42:42 crc kubenswrapper[4966]: I1217 08:42:42.521936 4966 generic.go:334] "Generic (PLEG): container finished" podID="b64018c5-5835-4e79-908f-b2f96d279dcf" containerID="0ea1617d5ee58d3f84abe10a555338f954881a61ba1214d724a9325aa57c958e" exitCode=0 Dec 17 08:42:42 crc kubenswrapper[4966]: I1217 08:42:42.522071 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" event={"ID":"b64018c5-5835-4e79-908f-b2f96d279dcf","Type":"ContainerDied","Data":"0ea1617d5ee58d3f84abe10a555338f954881a61ba1214d724a9325aa57c958e"} Dec 17 08:42:42 crc kubenswrapper[4966]: I1217 08:42:42.523060 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" event={"ID":"b64018c5-5835-4e79-908f-b2f96d279dcf","Type":"ContainerStarted","Data":"474d02e5919afcb5ed019add438df7306c2fcb8fceca9800115b204a79112a00"} Dec 17 08:42:42 crc kubenswrapper[4966]: I1217 08:42:42.921193 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:42 crc kubenswrapper[4966]: I1217 08:42:42.921746 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="ceilometer-central-agent" containerID="cri-o://cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f" gracePeriod=30 Dec 17 08:42:42 crc kubenswrapper[4966]: I1217 08:42:42.921887 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="ceilometer-notification-agent" containerID="cri-o://b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e" gracePeriod=30 Dec 17 08:42:42 crc kubenswrapper[4966]: I1217 08:42:42.921879 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="sg-core" containerID="cri-o://44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297" gracePeriod=30 Dec 17 08:42:42 crc kubenswrapper[4966]: I1217 08:42:42.922032 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="proxy-httpd" containerID="cri-o://473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780" gracePeriod=30 Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.404076 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.561811 4966 generic.go:334] "Generic (PLEG): container finished" podID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerID="473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780" exitCode=0 Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.561849 4966 generic.go:334] "Generic (PLEG): container finished" podID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerID="44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297" exitCode=2 Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.561859 4966 generic.go:334] "Generic (PLEG): container finished" podID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerID="cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f" exitCode=0 Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.561939 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerDied","Data":"473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780"} Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.561970 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerDied","Data":"44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297"} Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.561982 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerDied","Data":"cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f"} Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.569231 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" event={"ID":"b64018c5-5835-4e79-908f-b2f96d279dcf","Type":"ContainerStarted","Data":"762e174390ddcedf69063400ad72b0960633ed056a04985dbeaf65b3376401b8"} Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.569352 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-log" containerID="cri-o://9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a" gracePeriod=30 Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.569821 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-api" containerID="cri-o://0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f" gracePeriod=30 Dec 17 08:42:43 crc kubenswrapper[4966]: I1217 08:42:43.609613 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" podStartSLOduration=3.609598279 podStartE2EDuration="3.609598279s" podCreationTimestamp="2025-12-17 08:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:43.608755866 +0000 UTC m=+1299.153825818" watchObservedRunningTime="2025-12-17 08:42:43.609598279 +0000 UTC m=+1299.154668221" Dec 17 08:42:44 crc kubenswrapper[4966]: I1217 08:42:44.579945 4966 generic.go:334] "Generic (PLEG): container finished" podID="9014cabb-e109-4aed-99c8-990896bd10bf" containerID="9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a" exitCode=143 Dec 17 08:42:44 crc kubenswrapper[4966]: I1217 08:42:44.580034 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9014cabb-e109-4aed-99c8-990896bd10bf","Type":"ContainerDied","Data":"9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a"} Dec 17 08:42:44 crc kubenswrapper[4966]: I1217 08:42:44.580474 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.553679 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.595861 4966 generic.go:334] "Generic (PLEG): container finished" podID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerID="b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e" exitCode=0 Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.596946 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.597391 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerDied","Data":"b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e"} Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.597426 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4f6238b-6948-43ae-8a0e-1279ec490a71","Type":"ContainerDied","Data":"077aaaf8f2c5805c1ece5b9bd35d8a817ef02768838c39f6e78dab4442634eaf"} Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.597441 4966 scope.go:117] "RemoveContainer" containerID="473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.628669 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-log-httpd\") pod \"d4f6238b-6948-43ae-8a0e-1279ec490a71\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.628716 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-sg-core-conf-yaml\") pod \"d4f6238b-6948-43ae-8a0e-1279ec490a71\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.628818 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-run-httpd\") pod \"d4f6238b-6948-43ae-8a0e-1279ec490a71\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.629245 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d4f6238b-6948-43ae-8a0e-1279ec490a71" (UID: "d4f6238b-6948-43ae-8a0e-1279ec490a71"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.629296 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-combined-ca-bundle\") pod \"d4f6238b-6948-43ae-8a0e-1279ec490a71\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.629391 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d4f6238b-6948-43ae-8a0e-1279ec490a71" (UID: "d4f6238b-6948-43ae-8a0e-1279ec490a71"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.629710 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwfq9\" (UniqueName: \"kubernetes.io/projected/d4f6238b-6948-43ae-8a0e-1279ec490a71-kube-api-access-qwfq9\") pod \"d4f6238b-6948-43ae-8a0e-1279ec490a71\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.629750 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-config-data\") pod \"d4f6238b-6948-43ae-8a0e-1279ec490a71\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.629775 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-scripts\") pod \"d4f6238b-6948-43ae-8a0e-1279ec490a71\" (UID: \"d4f6238b-6948-43ae-8a0e-1279ec490a71\") " Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.630154 4966 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.630170 4966 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4f6238b-6948-43ae-8a0e-1279ec490a71-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.663796 4966 scope.go:117] "RemoveContainer" containerID="44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.668839 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-scripts" (OuterVolumeSpecName: "scripts") pod "d4f6238b-6948-43ae-8a0e-1279ec490a71" (UID: "d4f6238b-6948-43ae-8a0e-1279ec490a71"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.677731 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4f6238b-6948-43ae-8a0e-1279ec490a71-kube-api-access-qwfq9" (OuterVolumeSpecName: "kube-api-access-qwfq9") pod "d4f6238b-6948-43ae-8a0e-1279ec490a71" (UID: "d4f6238b-6948-43ae-8a0e-1279ec490a71"). InnerVolumeSpecName "kube-api-access-qwfq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.687025 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d4f6238b-6948-43ae-8a0e-1279ec490a71" (UID: "d4f6238b-6948-43ae-8a0e-1279ec490a71"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.734604 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.734630 4966 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.734642 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwfq9\" (UniqueName: \"kubernetes.io/projected/d4f6238b-6948-43ae-8a0e-1279ec490a71-kube-api-access-qwfq9\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.770016 4966 scope.go:117] "RemoveContainer" containerID="b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.798070 4966 scope.go:117] "RemoveContainer" containerID="cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.820013 4966 scope.go:117] "RemoveContainer" containerID="473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780" Dec 17 08:42:45 crc kubenswrapper[4966]: E1217 08:42:45.820363 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780\": container with ID starting with 473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780 not found: ID does not exist" containerID="473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.820404 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780"} err="failed to get container status \"473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780\": rpc error: code = NotFound desc = could not find container \"473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780\": container with ID starting with 473dcf76d89c82d13829794a19239f1985493c38f97f66a178c663dd94eb8780 not found: ID does not exist" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.820429 4966 scope.go:117] "RemoveContainer" containerID="44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297" Dec 17 08:42:45 crc kubenswrapper[4966]: E1217 08:42:45.820656 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297\": container with ID starting with 44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297 not found: ID does not exist" containerID="44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.820693 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297"} err="failed to get container status \"44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297\": rpc error: code = NotFound desc = could not find container \"44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297\": container with ID starting with 44839e26fab93b43c0aea7b8fed0ab35a84841d0e636195676f63091cc2f5297 not found: ID does not exist" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.820707 4966 scope.go:117] "RemoveContainer" containerID="b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e" Dec 17 08:42:45 crc kubenswrapper[4966]: E1217 08:42:45.820933 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e\": container with ID starting with b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e not found: ID does not exist" containerID="b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.820950 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e"} err="failed to get container status \"b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e\": rpc error: code = NotFound desc = could not find container \"b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e\": container with ID starting with b15b1b3c9ff0fe05e86607d5c7b6f33f4a266ddcaa7f4a1c1e22dbf6ef6f5f8e not found: ID does not exist" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.820961 4966 scope.go:117] "RemoveContainer" containerID="cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f" Dec 17 08:42:45 crc kubenswrapper[4966]: E1217 08:42:45.821139 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f\": container with ID starting with cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f not found: ID does not exist" containerID="cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.821152 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f"} err="failed to get container status \"cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f\": rpc error: code = NotFound desc = could not find container \"cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f\": container with ID starting with cde52142a494909a857d1c0caa2be8ebfdf69878ff93a4334ffc5db40132776f not found: ID does not exist" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.821690 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4f6238b-6948-43ae-8a0e-1279ec490a71" (UID: "d4f6238b-6948-43ae-8a0e-1279ec490a71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.828562 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-config-data" (OuterVolumeSpecName: "config-data") pod "d4f6238b-6948-43ae-8a0e-1279ec490a71" (UID: "d4f6238b-6948-43ae-8a0e-1279ec490a71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.836269 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.836322 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4f6238b-6948-43ae-8a0e-1279ec490a71-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.947784 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.960461 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.973456 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:45 crc kubenswrapper[4966]: E1217 08:42:45.974280 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="sg-core" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.974332 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="sg-core" Dec 17 08:42:45 crc kubenswrapper[4966]: E1217 08:42:45.974357 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="ceilometer-central-agent" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.974371 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="ceilometer-central-agent" Dec 17 08:42:45 crc kubenswrapper[4966]: E1217 08:42:45.974408 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="proxy-httpd" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.974420 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="proxy-httpd" Dec 17 08:42:45 crc kubenswrapper[4966]: E1217 08:42:45.974433 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="ceilometer-notification-agent" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.974444 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="ceilometer-notification-agent" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.974777 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="sg-core" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.974807 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="proxy-httpd" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.974845 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="ceilometer-notification-agent" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.974883 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" containerName="ceilometer-central-agent" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.977739 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.980791 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.981057 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 17 08:42:45 crc kubenswrapper[4966]: I1217 08:42:45.993983 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.041514 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zckj2\" (UniqueName: \"kubernetes.io/projected/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-kube-api-access-zckj2\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.041619 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-log-httpd\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.041651 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-run-httpd\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.041681 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.041701 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-config-data\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.041717 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-scripts\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.041735 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.143318 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zckj2\" (UniqueName: \"kubernetes.io/projected/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-kube-api-access-zckj2\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.143597 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-log-httpd\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.143736 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-run-httpd\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.143834 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.143937 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-config-data\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.144135 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-scripts\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.144235 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.144145 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-log-httpd\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.144392 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-run-httpd\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.150973 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-config-data\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.157185 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.159841 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-scripts\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.160313 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.166825 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zckj2\" (UniqueName: \"kubernetes.io/projected/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-kube-api-access-zckj2\") pod \"ceilometer-0\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.295273 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.927698 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4f6238b-6948-43ae-8a0e-1279ec490a71" path="/var/lib/kubelet/pods/d4f6238b-6948-43ae-8a0e-1279ec490a71/volumes" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.929245 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.929281 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:46 crc kubenswrapper[4966]: I1217 08:42:46.969595 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.425359 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.504348 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-config-data\") pod \"9014cabb-e109-4aed-99c8-990896bd10bf\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.504423 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-combined-ca-bundle\") pod \"9014cabb-e109-4aed-99c8-990896bd10bf\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.504849 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfzt8\" (UniqueName: \"kubernetes.io/projected/9014cabb-e109-4aed-99c8-990896bd10bf-kube-api-access-nfzt8\") pod \"9014cabb-e109-4aed-99c8-990896bd10bf\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.504903 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9014cabb-e109-4aed-99c8-990896bd10bf-logs\") pod \"9014cabb-e109-4aed-99c8-990896bd10bf\" (UID: \"9014cabb-e109-4aed-99c8-990896bd10bf\") " Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.506067 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9014cabb-e109-4aed-99c8-990896bd10bf-logs" (OuterVolumeSpecName: "logs") pod "9014cabb-e109-4aed-99c8-990896bd10bf" (UID: "9014cabb-e109-4aed-99c8-990896bd10bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.540269 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9014cabb-e109-4aed-99c8-990896bd10bf-kube-api-access-nfzt8" (OuterVolumeSpecName: "kube-api-access-nfzt8") pod "9014cabb-e109-4aed-99c8-990896bd10bf" (UID: "9014cabb-e109-4aed-99c8-990896bd10bf"). InnerVolumeSpecName "kube-api-access-nfzt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.570633 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-config-data" (OuterVolumeSpecName: "config-data") pod "9014cabb-e109-4aed-99c8-990896bd10bf" (UID: "9014cabb-e109-4aed-99c8-990896bd10bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.571182 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9014cabb-e109-4aed-99c8-990896bd10bf" (UID: "9014cabb-e109-4aed-99c8-990896bd10bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.609169 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.609192 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9014cabb-e109-4aed-99c8-990896bd10bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.609204 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfzt8\" (UniqueName: \"kubernetes.io/projected/9014cabb-e109-4aed-99c8-990896bd10bf-kube-api-access-nfzt8\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.609212 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9014cabb-e109-4aed-99c8-990896bd10bf-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.631640 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerStarted","Data":"3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe"} Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.631683 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerStarted","Data":"17c9460ed689ce578228515523787bf8613c37562e604f51e79fbb2d51fd2ef6"} Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.643398 4966 generic.go:334] "Generic (PLEG): container finished" podID="9014cabb-e109-4aed-99c8-990896bd10bf" containerID="0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f" exitCode=0 Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.644731 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.646993 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9014cabb-e109-4aed-99c8-990896bd10bf","Type":"ContainerDied","Data":"0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f"} Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.647022 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9014cabb-e109-4aed-99c8-990896bd10bf","Type":"ContainerDied","Data":"5ba6d85420f5904d4c2cc531c6e2bcfccbf55a1482a95165f89303be8d405cce"} Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.647037 4966 scope.go:117] "RemoveContainer" containerID="0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.698547 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.709671 4966 scope.go:117] "RemoveContainer" containerID="9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.735707 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.795412 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.798919 4966 scope.go:117] "RemoveContainer" containerID="0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f" Dec 17 08:42:47 crc kubenswrapper[4966]: E1217 08:42:47.817314 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f\": container with ID starting with 0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f not found: ID does not exist" containerID="0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.817367 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f"} err="failed to get container status \"0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f\": rpc error: code = NotFound desc = could not find container \"0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f\": container with ID starting with 0a002a180f7c89e65e21e0294b92f3c0dd6be6c426cbca9343cfbb1e6962407f not found: ID does not exist" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.817399 4966 scope.go:117] "RemoveContainer" containerID="9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.821316 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:47 crc kubenswrapper[4966]: E1217 08:42:47.821991 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a\": container with ID starting with 9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a not found: ID does not exist" containerID="9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.822044 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a"} err="failed to get container status \"9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a\": rpc error: code = NotFound desc = could not find container \"9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a\": container with ID starting with 9958525628dd09331fa8fad0f3cf2bc13590c3864aba78c06ba78188e22b181a not found: ID does not exist" Dec 17 08:42:47 crc kubenswrapper[4966]: E1217 08:42:47.823093 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-log" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.823113 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-log" Dec 17 08:42:47 crc kubenswrapper[4966]: E1217 08:42:47.823223 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-api" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.823230 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-api" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.825541 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-log" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.825669 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" containerName="nova-api-api" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.827341 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.832608 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.832973 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.837415 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.862208 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.923438 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.923654 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="bdca14dc-79c4-4e52-a748-9977656f7821" containerName="kube-state-metrics" containerID="cri-o://36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13" gracePeriod=30 Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.927503 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036fc8b2-4f95-496c-b364-29de6b452030-logs\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.927603 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-public-tls-certs\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.927955 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-internal-tls-certs\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.927988 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-config-data\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.928032 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zlch\" (UniqueName: \"kubernetes.io/projected/036fc8b2-4f95-496c-b364-29de6b452030-kube-api-access-4zlch\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.928067 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:47 crc kubenswrapper[4966]: I1217 08:42:47.980213 4966 scope.go:117] "RemoveContainer" containerID="9360b2c3dc58d21759bb96025da73cb17cd208ea4d9e5918c47555a440fccd99" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.005434 4966 scope.go:117] "RemoveContainer" containerID="efa8eb39af70f49a13d4f26ebb81dfe84ded7dca970249bd8725ad47b0b70921" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.033596 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036fc8b2-4f95-496c-b364-29de6b452030-logs\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.033653 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-public-tls-certs\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.034165 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036fc8b2-4f95-496c-b364-29de6b452030-logs\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.038837 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-internal-tls-certs\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.038890 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-config-data\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.038933 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zlch\" (UniqueName: \"kubernetes.io/projected/036fc8b2-4f95-496c-b364-29de6b452030-kube-api-access-4zlch\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.038977 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.039890 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-public-tls-certs\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.050280 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-config-data\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.050766 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-internal-tls-certs\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.068337 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.080184 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zlch\" (UniqueName: \"kubernetes.io/projected/036fc8b2-4f95-496c-b364-29de6b452030-kube-api-access-4zlch\") pod \"nova-api-0\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.097138 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-x4f2k"] Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.098723 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.102776 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.104438 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.120115 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-x4f2k"] Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.176197 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.253042 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-config-data\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.253085 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-scripts\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.253141 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxwlq\" (UniqueName: \"kubernetes.io/projected/4c08b24e-7324-49f1-b084-fd2f866dd7c4-kube-api-access-kxwlq\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.253456 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.366680 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-config-data\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.366837 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-scripts\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.367014 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxwlq\" (UniqueName: \"kubernetes.io/projected/4c08b24e-7324-49f1-b084-fd2f866dd7c4-kube-api-access-kxwlq\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.367815 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.373652 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-scripts\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.390799 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxwlq\" (UniqueName: \"kubernetes.io/projected/4c08b24e-7324-49f1-b084-fd2f866dd7c4-kube-api-access-kxwlq\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.398116 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-config-data\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.464027 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-x4f2k\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.633001 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.734194 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.742673 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerStarted","Data":"833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01"} Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.771220 4966 generic.go:334] "Generic (PLEG): container finished" podID="bdca14dc-79c4-4e52-a748-9977656f7821" containerID="36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13" exitCode=2 Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.772130 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bdca14dc-79c4-4e52-a748-9977656f7821","Type":"ContainerDied","Data":"36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13"} Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.772170 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bdca14dc-79c4-4e52-a748-9977656f7821","Type":"ContainerDied","Data":"f4e0b3e58bf71c2d2ef5bd2388e9ada27432f48abe12780d08689b80a5997f1c"} Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.772210 4966 scope.go:117] "RemoveContainer" containerID="36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.772214 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.787179 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2nbd\" (UniqueName: \"kubernetes.io/projected/bdca14dc-79c4-4e52-a748-9977656f7821-kube-api-access-t2nbd\") pod \"bdca14dc-79c4-4e52-a748-9977656f7821\" (UID: \"bdca14dc-79c4-4e52-a748-9977656f7821\") " Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.809629 4966 scope.go:117] "RemoveContainer" containerID="36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13" Dec 17 08:42:48 crc kubenswrapper[4966]: E1217 08:42:48.810355 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13\": container with ID starting with 36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13 not found: ID does not exist" containerID="36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.810491 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13"} err="failed to get container status \"36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13\": rpc error: code = NotFound desc = could not find container \"36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13\": container with ID starting with 36ec50e03dac3b4a3cf4be0b893e6ef12df5efbe430031fd6694f067d88bab13 not found: ID does not exist" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.827446 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdca14dc-79c4-4e52-a748-9977656f7821-kube-api-access-t2nbd" (OuterVolumeSpecName: "kube-api-access-t2nbd") pod "bdca14dc-79c4-4e52-a748-9977656f7821" (UID: "bdca14dc-79c4-4e52-a748-9977656f7821"). InnerVolumeSpecName "kube-api-access-t2nbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.888929 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9014cabb-e109-4aed-99c8-990896bd10bf" path="/var/lib/kubelet/pods/9014cabb-e109-4aed-99c8-990896bd10bf/volumes" Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.897915 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:48 crc kubenswrapper[4966]: I1217 08:42:48.899314 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2nbd\" (UniqueName: \"kubernetes.io/projected/bdca14dc-79c4-4e52-a748-9977656f7821-kube-api-access-t2nbd\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.106148 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.123670 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.142936 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 17 08:42:49 crc kubenswrapper[4966]: E1217 08:42:49.143825 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdca14dc-79c4-4e52-a748-9977656f7821" containerName="kube-state-metrics" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.143968 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdca14dc-79c4-4e52-a748-9977656f7821" containerName="kube-state-metrics" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.144359 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdca14dc-79c4-4e52-a748-9977656f7821" containerName="kube-state-metrics" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.145174 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.148627 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.150068 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.153983 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.320910 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d38a31c8-7333-40cb-b4ba-e7d172379f28-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.321235 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d38a31c8-7333-40cb-b4ba-e7d172379f28-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.321261 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a31c8-7333-40cb-b4ba-e7d172379f28-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.321281 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drzcr\" (UniqueName: \"kubernetes.io/projected/d38a31c8-7333-40cb-b4ba-e7d172379f28-kube-api-access-drzcr\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.419354 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-x4f2k"] Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.423113 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d38a31c8-7333-40cb-b4ba-e7d172379f28-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.423172 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d38a31c8-7333-40cb-b4ba-e7d172379f28-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.423198 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a31c8-7333-40cb-b4ba-e7d172379f28-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.423216 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drzcr\" (UniqueName: \"kubernetes.io/projected/d38a31c8-7333-40cb-b4ba-e7d172379f28-kube-api-access-drzcr\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.445719 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a31c8-7333-40cb-b4ba-e7d172379f28-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.446283 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d38a31c8-7333-40cb-b4ba-e7d172379f28-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.446811 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d38a31c8-7333-40cb-b4ba-e7d172379f28-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.453574 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drzcr\" (UniqueName: \"kubernetes.io/projected/d38a31c8-7333-40cb-b4ba-e7d172379f28-kube-api-access-drzcr\") pod \"kube-state-metrics-0\" (UID: \"d38a31c8-7333-40cb-b4ba-e7d172379f28\") " pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.601061 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.795195 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerStarted","Data":"d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1"} Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.806596 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x4f2k" event={"ID":"4c08b24e-7324-49f1-b084-fd2f866dd7c4","Type":"ContainerStarted","Data":"180fd7d42e953b5e6e39dcf5128828e9ee22666ca800f525554c59b6ea800a6c"} Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.813972 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036fc8b2-4f95-496c-b364-29de6b452030","Type":"ContainerStarted","Data":"3864b2899f39f6c2897d12bdbcc9d0e9465fb5e8a94aa91ec8a4b42a31468ef8"} Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.814030 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036fc8b2-4f95-496c-b364-29de6b452030","Type":"ContainerStarted","Data":"d700a688723c1a8be0871b4f375e112861798ee6d37ce059556167d96721dd43"} Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.814044 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036fc8b2-4f95-496c-b364-29de6b452030","Type":"ContainerStarted","Data":"c8722be9b9ea15b0196bb8beb6aafdc0cf09036a3c2f5dd481a89978c025b0d0"} Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.830817 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-x4f2k" podStartSLOduration=1.830800062 podStartE2EDuration="1.830800062s" podCreationTimestamp="2025-12-17 08:42:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:49.830158604 +0000 UTC m=+1305.375228546" watchObservedRunningTime="2025-12-17 08:42:49.830800062 +0000 UTC m=+1305.375870004" Dec 17 08:42:49 crc kubenswrapper[4966]: I1217 08:42:49.848349 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.848330133 podStartE2EDuration="2.848330133s" podCreationTimestamp="2025-12-17 08:42:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:42:49.847395667 +0000 UTC m=+1305.392465629" watchObservedRunningTime="2025-12-17 08:42:49.848330133 +0000 UTC m=+1305.393400075" Dec 17 08:42:50 crc kubenswrapper[4966]: I1217 08:42:50.201652 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 17 08:42:50 crc kubenswrapper[4966]: I1217 08:42:50.626035 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:50 crc kubenswrapper[4966]: I1217 08:42:50.824592 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d38a31c8-7333-40cb-b4ba-e7d172379f28","Type":"ContainerStarted","Data":"8ec7a646bccb45fdc2982f4549591371961025c90be7c5d8ba2cba87b25c30fe"} Dec 17 08:42:50 crc kubenswrapper[4966]: I1217 08:42:50.827245 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x4f2k" event={"ID":"4c08b24e-7324-49f1-b084-fd2f866dd7c4","Type":"ContainerStarted","Data":"ef731192a3d3c582c54662603c920c08d8faf1d504da7a9e068e49425bbaa852"} Dec 17 08:42:50 crc kubenswrapper[4966]: I1217 08:42:50.840710 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdca14dc-79c4-4e52-a748-9977656f7821" path="/var/lib/kubelet/pods/bdca14dc-79c4-4e52-a748-9977656f7821/volumes" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.090052 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.153586 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d88fdb6c9-zmf26"] Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.153821 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" podUID="06f02ab9-5bb8-47ba-88a8-9971cf378afa" containerName="dnsmasq-dns" containerID="cri-o://71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7" gracePeriod=10 Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.798938 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.846140 4966 generic.go:334] "Generic (PLEG): container finished" podID="06f02ab9-5bb8-47ba-88a8-9971cf378afa" containerID="71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7" exitCode=0 Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.846304 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.847162 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" event={"ID":"06f02ab9-5bb8-47ba-88a8-9971cf378afa","Type":"ContainerDied","Data":"71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7"} Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.847194 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d88fdb6c9-zmf26" event={"ID":"06f02ab9-5bb8-47ba-88a8-9971cf378afa","Type":"ContainerDied","Data":"11e9e33eddd638a04f5ea126c1248301b5b2a3eb572fd1553f33584ca1e46264"} Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.847217 4966 scope.go:117] "RemoveContainer" containerID="71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.883399 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d38a31c8-7333-40cb-b4ba-e7d172379f28","Type":"ContainerStarted","Data":"ce6f750174936aac29695f62bb86f74bb32308add38f1f75620cd04dee95304a"} Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.884384 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.915491 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.575039044 podStartE2EDuration="2.915476716s" podCreationTimestamp="2025-12-17 08:42:49 +0000 UTC" firstStartedPulling="2025-12-17 08:42:50.205246417 +0000 UTC m=+1305.750316359" lastFinishedPulling="2025-12-17 08:42:50.545684089 +0000 UTC m=+1306.090754031" observedRunningTime="2025-12-17 08:42:51.914027336 +0000 UTC m=+1307.459097298" watchObservedRunningTime="2025-12-17 08:42:51.915476716 +0000 UTC m=+1307.460546658" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.937717 4966 scope.go:117] "RemoveContainer" containerID="99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.960004 4966 scope.go:117] "RemoveContainer" containerID="71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7" Dec 17 08:42:51 crc kubenswrapper[4966]: E1217 08:42:51.960423 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7\": container with ID starting with 71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7 not found: ID does not exist" containerID="71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.960472 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7"} err="failed to get container status \"71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7\": rpc error: code = NotFound desc = could not find container \"71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7\": container with ID starting with 71047174e7ec396c6ab3a97563c82e226f8c24019b28a9625aa5eac52d7e95b7 not found: ID does not exist" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.960506 4966 scope.go:117] "RemoveContainer" containerID="99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51" Dec 17 08:42:51 crc kubenswrapper[4966]: E1217 08:42:51.961068 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51\": container with ID starting with 99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51 not found: ID does not exist" containerID="99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.961107 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51"} err="failed to get container status \"99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51\": rpc error: code = NotFound desc = could not find container \"99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51\": container with ID starting with 99a77db3e9f94b8d678a6e4cdfdc3e7e3f465f8b7874b8d9dca47fb35fd45a51 not found: ID does not exist" Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.998639 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4pds\" (UniqueName: \"kubernetes.io/projected/06f02ab9-5bb8-47ba-88a8-9971cf378afa-kube-api-access-t4pds\") pod \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.998771 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-nb\") pod \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.998834 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-sb\") pod \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.998857 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-config\") pod \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.998933 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-svc\") pod \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " Dec 17 08:42:51 crc kubenswrapper[4966]: I1217 08:42:51.999020 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-swift-storage-0\") pod \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\" (UID: \"06f02ab9-5bb8-47ba-88a8-9971cf378afa\") " Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.012064 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f02ab9-5bb8-47ba-88a8-9971cf378afa-kube-api-access-t4pds" (OuterVolumeSpecName: "kube-api-access-t4pds") pod "06f02ab9-5bb8-47ba-88a8-9971cf378afa" (UID: "06f02ab9-5bb8-47ba-88a8-9971cf378afa"). InnerVolumeSpecName "kube-api-access-t4pds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.085663 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-config" (OuterVolumeSpecName: "config") pod "06f02ab9-5bb8-47ba-88a8-9971cf378afa" (UID: "06f02ab9-5bb8-47ba-88a8-9971cf378afa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.086780 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "06f02ab9-5bb8-47ba-88a8-9971cf378afa" (UID: "06f02ab9-5bb8-47ba-88a8-9971cf378afa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.088382 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "06f02ab9-5bb8-47ba-88a8-9971cf378afa" (UID: "06f02ab9-5bb8-47ba-88a8-9971cf378afa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.091732 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "06f02ab9-5bb8-47ba-88a8-9971cf378afa" (UID: "06f02ab9-5bb8-47ba-88a8-9971cf378afa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.098464 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "06f02ab9-5bb8-47ba-88a8-9971cf378afa" (UID: "06f02ab9-5bb8-47ba-88a8-9971cf378afa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.101844 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.101900 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4pds\" (UniqueName: \"kubernetes.io/projected/06f02ab9-5bb8-47ba-88a8-9971cf378afa-kube-api-access-t4pds\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.101916 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.101925 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.101934 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.101942 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06f02ab9-5bb8-47ba-88a8-9971cf378afa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.177631 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d88fdb6c9-zmf26"] Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.192396 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d88fdb6c9-zmf26"] Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.840682 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f02ab9-5bb8-47ba-88a8-9971cf378afa" path="/var/lib/kubelet/pods/06f02ab9-5bb8-47ba-88a8-9971cf378afa/volumes" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.897339 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerStarted","Data":"83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305"} Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.897472 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="ceilometer-central-agent" containerID="cri-o://3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe" gracePeriod=30 Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.897559 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="ceilometer-notification-agent" containerID="cri-o://833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01" gracePeriod=30 Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.897559 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="sg-core" containerID="cri-o://d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1" gracePeriod=30 Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.897633 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="proxy-httpd" containerID="cri-o://83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305" gracePeriod=30 Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.897686 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 17 08:42:52 crc kubenswrapper[4966]: I1217 08:42:52.924205 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.3158184410000002 podStartE2EDuration="7.924185926s" podCreationTimestamp="2025-12-17 08:42:45 +0000 UTC" firstStartedPulling="2025-12-17 08:42:46.926297861 +0000 UTC m=+1302.471367803" lastFinishedPulling="2025-12-17 08:42:51.534665346 +0000 UTC m=+1307.079735288" observedRunningTime="2025-12-17 08:42:52.916294259 +0000 UTC m=+1308.461364201" watchObservedRunningTime="2025-12-17 08:42:52.924185926 +0000 UTC m=+1308.469255868" Dec 17 08:42:53 crc kubenswrapper[4966]: I1217 08:42:53.909052 4966 generic.go:334] "Generic (PLEG): container finished" podID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerID="83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305" exitCode=0 Dec 17 08:42:53 crc kubenswrapper[4966]: I1217 08:42:53.909276 4966 generic.go:334] "Generic (PLEG): container finished" podID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerID="d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1" exitCode=2 Dec 17 08:42:53 crc kubenswrapper[4966]: I1217 08:42:53.909284 4966 generic.go:334] "Generic (PLEG): container finished" podID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerID="833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01" exitCode=0 Dec 17 08:42:53 crc kubenswrapper[4966]: I1217 08:42:53.909115 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerDied","Data":"83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305"} Dec 17 08:42:53 crc kubenswrapper[4966]: I1217 08:42:53.909313 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerDied","Data":"d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1"} Dec 17 08:42:53 crc kubenswrapper[4966]: I1217 08:42:53.909322 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerDied","Data":"833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01"} Dec 17 08:42:56 crc kubenswrapper[4966]: I1217 08:42:56.936837 4966 generic.go:334] "Generic (PLEG): container finished" podID="4c08b24e-7324-49f1-b084-fd2f866dd7c4" containerID="ef731192a3d3c582c54662603c920c08d8faf1d504da7a9e068e49425bbaa852" exitCode=0 Dec 17 08:42:56 crc kubenswrapper[4966]: I1217 08:42:56.936919 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x4f2k" event={"ID":"4c08b24e-7324-49f1-b084-fd2f866dd7c4","Type":"ContainerDied","Data":"ef731192a3d3c582c54662603c920c08d8faf1d504da7a9e068e49425bbaa852"} Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.476930 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.606512 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-scripts\") pod \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.606582 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-sg-core-conf-yaml\") pod \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.606640 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-log-httpd\") pod \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.606694 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-combined-ca-bundle\") pod \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.606748 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-config-data\") pod \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.606825 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zckj2\" (UniqueName: \"kubernetes.io/projected/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-kube-api-access-zckj2\") pod \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.606918 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-run-httpd\") pod \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\" (UID: \"91e1ea08-43e9-46b7-aa6c-17bb16bc3079\") " Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.607391 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "91e1ea08-43e9-46b7-aa6c-17bb16bc3079" (UID: "91e1ea08-43e9-46b7-aa6c-17bb16bc3079"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.607538 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "91e1ea08-43e9-46b7-aa6c-17bb16bc3079" (UID: "91e1ea08-43e9-46b7-aa6c-17bb16bc3079"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.620085 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-kube-api-access-zckj2" (OuterVolumeSpecName: "kube-api-access-zckj2") pod "91e1ea08-43e9-46b7-aa6c-17bb16bc3079" (UID: "91e1ea08-43e9-46b7-aa6c-17bb16bc3079"). InnerVolumeSpecName "kube-api-access-zckj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.626327 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-scripts" (OuterVolumeSpecName: "scripts") pod "91e1ea08-43e9-46b7-aa6c-17bb16bc3079" (UID: "91e1ea08-43e9-46b7-aa6c-17bb16bc3079"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.656183 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "91e1ea08-43e9-46b7-aa6c-17bb16bc3079" (UID: "91e1ea08-43e9-46b7-aa6c-17bb16bc3079"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.710261 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zckj2\" (UniqueName: \"kubernetes.io/projected/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-kube-api-access-zckj2\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.710301 4966 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.710310 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.710319 4966 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.710327 4966 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.822112 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-config-data" (OuterVolumeSpecName: "config-data") pod "91e1ea08-43e9-46b7-aa6c-17bb16bc3079" (UID: "91e1ea08-43e9-46b7-aa6c-17bb16bc3079"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.822197 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91e1ea08-43e9-46b7-aa6c-17bb16bc3079" (UID: "91e1ea08-43e9-46b7-aa6c-17bb16bc3079"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.914568 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.914604 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91e1ea08-43e9-46b7-aa6c-17bb16bc3079-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.948453 4966 generic.go:334] "Generic (PLEG): container finished" podID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerID="3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe" exitCode=0 Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.948682 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.951042 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerDied","Data":"3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe"} Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.951096 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"91e1ea08-43e9-46b7-aa6c-17bb16bc3079","Type":"ContainerDied","Data":"17c9460ed689ce578228515523787bf8613c37562e604f51e79fbb2d51fd2ef6"} Dec 17 08:42:57 crc kubenswrapper[4966]: I1217 08:42:57.951117 4966 scope.go:117] "RemoveContainer" containerID="83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.006035 4966 scope.go:117] "RemoveContainer" containerID="d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.021599 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.029604 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91e1ea08_43e9_46b7_aa6c_17bb16bc3079.slice\": RecentStats: unable to find data in memory cache]" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.048763 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.077196 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.078270 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f02ab9-5bb8-47ba-88a8-9971cf378afa" containerName="dnsmasq-dns" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.078353 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f02ab9-5bb8-47ba-88a8-9971cf378afa" containerName="dnsmasq-dns" Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.079342 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="ceilometer-central-agent" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.079434 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="ceilometer-central-agent" Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.079508 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="sg-core" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.079571 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="sg-core" Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.079644 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f02ab9-5bb8-47ba-88a8-9971cf378afa" containerName="init" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.079732 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f02ab9-5bb8-47ba-88a8-9971cf378afa" containerName="init" Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.079821 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="proxy-httpd" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.079943 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="proxy-httpd" Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.080034 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="ceilometer-notification-agent" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.080096 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="ceilometer-notification-agent" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.080660 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f02ab9-5bb8-47ba-88a8-9971cf378afa" containerName="dnsmasq-dns" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.080758 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="proxy-httpd" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.080839 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="ceilometer-central-agent" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.080955 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="sg-core" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.081063 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" containerName="ceilometer-notification-agent" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.084515 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.084723 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.089234 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.090291 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.090706 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.097779 4966 scope.go:117] "RemoveContainer" containerID="833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.175023 4966 scope.go:117] "RemoveContainer" containerID="3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.177537 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.177632 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.210752 4966 scope.go:117] "RemoveContainer" containerID="83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305" Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.211491 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305\": container with ID starting with 83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305 not found: ID does not exist" containerID="83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.211556 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305"} err="failed to get container status \"83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305\": rpc error: code = NotFound desc = could not find container \"83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305\": container with ID starting with 83c2d0055117833ca2df7298e74529d56409bab844e3ce472e2e5e5a5dd59305 not found: ID does not exist" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.211639 4966 scope.go:117] "RemoveContainer" containerID="d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1" Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.212514 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1\": container with ID starting with d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1 not found: ID does not exist" containerID="d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.212542 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1"} err="failed to get container status \"d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1\": rpc error: code = NotFound desc = could not find container \"d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1\": container with ID starting with d4ae187d469bfaee2e7b8dcda6dccc2e71e4e9d70f6f7dd5dae53283deddcae1 not found: ID does not exist" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.212557 4966 scope.go:117] "RemoveContainer" containerID="833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01" Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.213260 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01\": container with ID starting with 833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01 not found: ID does not exist" containerID="833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.213281 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01"} err="failed to get container status \"833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01\": rpc error: code = NotFound desc = could not find container \"833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01\": container with ID starting with 833f98e1d0f60578337c7ec623b3a3004e352d3f70e4443b24ad10590b18aa01 not found: ID does not exist" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.213293 4966 scope.go:117] "RemoveContainer" containerID="3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe" Dec 17 08:42:58 crc kubenswrapper[4966]: E1217 08:42:58.214134 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe\": container with ID starting with 3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe not found: ID does not exist" containerID="3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.214175 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe"} err="failed to get container status \"3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe\": rpc error: code = NotFound desc = could not find container \"3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe\": container with ID starting with 3806754cc90ccb12d2a347c0a60dd34a07799fbf74247532e60e1abee9bc8cbe not found: ID does not exist" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.223825 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-scripts\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.223948 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2x2q\" (UniqueName: \"kubernetes.io/projected/bd0eecf1-7a58-4d65-b7db-7fedc658de06-kube-api-access-j2x2q\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.223970 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.223999 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd0eecf1-7a58-4d65-b7db-7fedc658de06-log-httpd\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.224049 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd0eecf1-7a58-4d65-b7db-7fedc658de06-run-httpd\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.224067 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.224087 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-config-data\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.224107 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.326462 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2x2q\" (UniqueName: \"kubernetes.io/projected/bd0eecf1-7a58-4d65-b7db-7fedc658de06-kube-api-access-j2x2q\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.326722 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.326765 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd0eecf1-7a58-4d65-b7db-7fedc658de06-log-httpd\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.326808 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd0eecf1-7a58-4d65-b7db-7fedc658de06-run-httpd\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.326824 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.326844 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-config-data\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.327880 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.328036 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-scripts\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.328249 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd0eecf1-7a58-4d65-b7db-7fedc658de06-log-httpd\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.328778 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd0eecf1-7a58-4d65-b7db-7fedc658de06-run-httpd\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.333430 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.346717 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.353597 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2x2q\" (UniqueName: \"kubernetes.io/projected/bd0eecf1-7a58-4d65-b7db-7fedc658de06-kube-api-access-j2x2q\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.354529 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-scripts\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.368014 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-config-data\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.368661 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd0eecf1-7a58-4d65-b7db-7fedc658de06-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd0eecf1-7a58-4d65-b7db-7fedc658de06\") " pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.409844 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.542673 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.639939 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-scripts\") pod \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.640083 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-combined-ca-bundle\") pod \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.640151 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-config-data\") pod \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.640256 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxwlq\" (UniqueName: \"kubernetes.io/projected/4c08b24e-7324-49f1-b084-fd2f866dd7c4-kube-api-access-kxwlq\") pod \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\" (UID: \"4c08b24e-7324-49f1-b084-fd2f866dd7c4\") " Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.644108 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c08b24e-7324-49f1-b084-fd2f866dd7c4-kube-api-access-kxwlq" (OuterVolumeSpecName: "kube-api-access-kxwlq") pod "4c08b24e-7324-49f1-b084-fd2f866dd7c4" (UID: "4c08b24e-7324-49f1-b084-fd2f866dd7c4"). InnerVolumeSpecName "kube-api-access-kxwlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.658851 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-scripts" (OuterVolumeSpecName: "scripts") pod "4c08b24e-7324-49f1-b084-fd2f866dd7c4" (UID: "4c08b24e-7324-49f1-b084-fd2f866dd7c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.712599 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c08b24e-7324-49f1-b084-fd2f866dd7c4" (UID: "4c08b24e-7324-49f1-b084-fd2f866dd7c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.719250 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-config-data" (OuterVolumeSpecName: "config-data") pod "4c08b24e-7324-49f1-b084-fd2f866dd7c4" (UID: "4c08b24e-7324-49f1-b084-fd2f866dd7c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.742002 4966 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-scripts\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.742035 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.742050 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c08b24e-7324-49f1-b084-fd2f866dd7c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.742061 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxwlq\" (UniqueName: \"kubernetes.io/projected/4c08b24e-7324-49f1-b084-fd2f866dd7c4-kube-api-access-kxwlq\") on node \"crc\" DevicePath \"\"" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.840716 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91e1ea08-43e9-46b7-aa6c-17bb16bc3079" path="/var/lib/kubelet/pods/91e1ea08-43e9-46b7-aa6c-17bb16bc3079/volumes" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.947079 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.963273 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-x4f2k" event={"ID":"4c08b24e-7324-49f1-b084-fd2f866dd7c4","Type":"ContainerDied","Data":"180fd7d42e953b5e6e39dcf5128828e9ee22666ca800f525554c59b6ea800a6c"} Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.963310 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="180fd7d42e953b5e6e39dcf5128828e9ee22666ca800f525554c59b6ea800a6c" Dec 17 08:42:58 crc kubenswrapper[4966]: I1217 08:42:58.963365 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-x4f2k" Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.157360 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.157940 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-log" containerID="cri-o://d700a688723c1a8be0871b4f375e112861798ee6d37ce059556167d96721dd43" gracePeriod=30 Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.158028 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-api" containerID="cri-o://3864b2899f39f6c2897d12bdbcc9d0e9465fb5e8a94aa91ec8a4b42a31468ef8" gracePeriod=30 Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.171745 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.218:8774/\": EOF" Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.175446 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.175633 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="03b5489a-56c8-441a-8fb6-421ac6275281" containerName="nova-scheduler-scheduler" containerID="cri-o://acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" gracePeriod=30 Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.196476 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.196707 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-log" containerID="cri-o://035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32" gracePeriod=30 Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.196746 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-metadata" containerID="cri-o://048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e" gracePeriod=30 Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.210079 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.218:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.627810 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 17 08:42:59 crc kubenswrapper[4966]: E1217 08:42:59.713517 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 17 08:42:59 crc kubenswrapper[4966]: E1217 08:42:59.720445 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 17 08:42:59 crc kubenswrapper[4966]: E1217 08:42:59.722495 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 17 08:42:59 crc kubenswrapper[4966]: E1217 08:42:59.722574 4966 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="03b5489a-56c8-441a-8fb6-421ac6275281" containerName="nova-scheduler-scheduler" Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.973917 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd0eecf1-7a58-4d65-b7db-7fedc658de06","Type":"ContainerStarted","Data":"4a46eb62a02cf42b51691fb79b6aedff570c8119a3c82a0d2b0fefda930ee526"} Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.973968 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd0eecf1-7a58-4d65-b7db-7fedc658de06","Type":"ContainerStarted","Data":"9fbc1799bc73403527342437bff6ba9644c7f0ab32a82bcbde87e773a51a7f54"} Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.973978 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd0eecf1-7a58-4d65-b7db-7fedc658de06","Type":"ContainerStarted","Data":"e09c2df70e74266f42f4b56f2fbdf70e27f99af3262cd32d5057d1aa1953783f"} Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.976160 4966 generic.go:334] "Generic (PLEG): container finished" podID="036fc8b2-4f95-496c-b364-29de6b452030" containerID="d700a688723c1a8be0871b4f375e112861798ee6d37ce059556167d96721dd43" exitCode=143 Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.976235 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036fc8b2-4f95-496c-b364-29de6b452030","Type":"ContainerDied","Data":"d700a688723c1a8be0871b4f375e112861798ee6d37ce059556167d96721dd43"} Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.978846 4966 generic.go:334] "Generic (PLEG): container finished" podID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerID="035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32" exitCode=143 Dec 17 08:42:59 crc kubenswrapper[4966]: I1217 08:42:59.978898 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d3b5f2dd-052a-4bb8-afaf-96b9ef596219","Type":"ContainerDied","Data":"035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32"} Dec 17 08:43:01 crc kubenswrapper[4966]: I1217 08:43:01.003458 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd0eecf1-7a58-4d65-b7db-7fedc658de06","Type":"ContainerStarted","Data":"e44c0690a40a7c8e34eb907a1afb267754e300ccdff8334ca22f33a86498e2b6"} Dec 17 08:43:02 crc kubenswrapper[4966]: I1217 08:43:02.889307 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:43:02 crc kubenswrapper[4966]: I1217 08:43:02.939938 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-combined-ca-bundle\") pod \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " Dec 17 08:43:02 crc kubenswrapper[4966]: I1217 08:43:02.940028 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-config-data\") pod \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " Dec 17 08:43:02 crc kubenswrapper[4966]: I1217 08:43:02.940123 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-nova-metadata-tls-certs\") pod \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " Dec 17 08:43:02 crc kubenswrapper[4966]: I1217 08:43:02.940161 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjbwx\" (UniqueName: \"kubernetes.io/projected/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-kube-api-access-sjbwx\") pod \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " Dec 17 08:43:02 crc kubenswrapper[4966]: I1217 08:43:02.940263 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-logs\") pod \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\" (UID: \"d3b5f2dd-052a-4bb8-afaf-96b9ef596219\") " Dec 17 08:43:02 crc kubenswrapper[4966]: I1217 08:43:02.941288 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-logs" (OuterVolumeSpecName: "logs") pod "d3b5f2dd-052a-4bb8-afaf-96b9ef596219" (UID: "d3b5f2dd-052a-4bb8-afaf-96b9ef596219"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:43:02 crc kubenswrapper[4966]: I1217 08:43:02.947980 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-kube-api-access-sjbwx" (OuterVolumeSpecName: "kube-api-access-sjbwx") pod "d3b5f2dd-052a-4bb8-afaf-96b9ef596219" (UID: "d3b5f2dd-052a-4bb8-afaf-96b9ef596219"). InnerVolumeSpecName "kube-api-access-sjbwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.029264 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-config-data" (OuterVolumeSpecName: "config-data") pod "d3b5f2dd-052a-4bb8-afaf-96b9ef596219" (UID: "d3b5f2dd-052a-4bb8-afaf-96b9ef596219"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.043319 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.043355 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjbwx\" (UniqueName: \"kubernetes.io/projected/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-kube-api-access-sjbwx\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.043366 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.057642 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd0eecf1-7a58-4d65-b7db-7fedc658de06","Type":"ContainerStarted","Data":"4b98df18ca366e68e583bbc817a15ccd330b2b60bc11154fc7f9966a1579f4c4"} Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.058080 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.061580 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3b5f2dd-052a-4bb8-afaf-96b9ef596219" (UID: "d3b5f2dd-052a-4bb8-afaf-96b9ef596219"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.064108 4966 generic.go:334] "Generic (PLEG): container finished" podID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerID="048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e" exitCode=0 Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.064158 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d3b5f2dd-052a-4bb8-afaf-96b9ef596219","Type":"ContainerDied","Data":"048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e"} Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.064202 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d3b5f2dd-052a-4bb8-afaf-96b9ef596219","Type":"ContainerDied","Data":"4ea51ce368438542045aea8f1dd61eff8fd9f4bfc65b5d2380d87ee3ddb49ee6"} Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.064222 4966 scope.go:117] "RemoveContainer" containerID="048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.064354 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.085102 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d3b5f2dd-052a-4bb8-afaf-96b9ef596219" (UID: "d3b5f2dd-052a-4bb8-afaf-96b9ef596219"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.116096 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.164070133 podStartE2EDuration="5.116074287s" podCreationTimestamp="2025-12-17 08:42:58 +0000 UTC" firstStartedPulling="2025-12-17 08:42:58.955064736 +0000 UTC m=+1314.500134678" lastFinishedPulling="2025-12-17 08:43:01.90706889 +0000 UTC m=+1317.452138832" observedRunningTime="2025-12-17 08:43:03.106290028 +0000 UTC m=+1318.651359980" watchObservedRunningTime="2025-12-17 08:43:03.116074287 +0000 UTC m=+1318.661144249" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.145758 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.145788 4966 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b5f2dd-052a-4bb8-afaf-96b9ef596219-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.163685 4966 scope.go:117] "RemoveContainer" containerID="035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.192369 4966 scope.go:117] "RemoveContainer" containerID="048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e" Dec 17 08:43:03 crc kubenswrapper[4966]: E1217 08:43:03.192857 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e\": container with ID starting with 048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e not found: ID does not exist" containerID="048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.192937 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e"} err="failed to get container status \"048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e\": rpc error: code = NotFound desc = could not find container \"048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e\": container with ID starting with 048869c505499c33f53c0858b2ca529aff83000e1f41d09a720679ba843fd26e not found: ID does not exist" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.192997 4966 scope.go:117] "RemoveContainer" containerID="035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32" Dec 17 08:43:03 crc kubenswrapper[4966]: E1217 08:43:03.193349 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32\": container with ID starting with 035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32 not found: ID does not exist" containerID="035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.193377 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32"} err="failed to get container status \"035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32\": rpc error: code = NotFound desc = could not find container \"035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32\": container with ID starting with 035c373c736ceadfe0205649b4fee23c1079c900e651179687f5dabb8f595d32 not found: ID does not exist" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.402776 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.413619 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.429941 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:43:03 crc kubenswrapper[4966]: E1217 08:43:03.430318 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c08b24e-7324-49f1-b084-fd2f866dd7c4" containerName="nova-manage" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.430338 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c08b24e-7324-49f1-b084-fd2f866dd7c4" containerName="nova-manage" Dec 17 08:43:03 crc kubenswrapper[4966]: E1217 08:43:03.430380 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-metadata" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.430387 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-metadata" Dec 17 08:43:03 crc kubenswrapper[4966]: E1217 08:43:03.430402 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-log" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.430408 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-log" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.430586 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-metadata" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.430605 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c08b24e-7324-49f1-b084-fd2f866dd7c4" containerName="nova-manage" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.430622 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-log" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.431573 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.434250 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.436050 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.447310 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.450072 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6f61a35-fecb-4c95-b487-94dcfc651311-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.450440 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6f61a35-fecb-4c95-b487-94dcfc651311-logs\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.450669 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6f61a35-fecb-4c95-b487-94dcfc651311-config-data\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.450774 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6rgf\" (UniqueName: \"kubernetes.io/projected/a6f61a35-fecb-4c95-b487-94dcfc651311-kube-api-access-m6rgf\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.450827 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6f61a35-fecb-4c95-b487-94dcfc651311-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.552289 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6f61a35-fecb-4c95-b487-94dcfc651311-config-data\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.552347 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6rgf\" (UniqueName: \"kubernetes.io/projected/a6f61a35-fecb-4c95-b487-94dcfc651311-kube-api-access-m6rgf\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.552368 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6f61a35-fecb-4c95-b487-94dcfc651311-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.552439 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6f61a35-fecb-4c95-b487-94dcfc651311-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.552511 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6f61a35-fecb-4c95-b487-94dcfc651311-logs\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.552951 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6f61a35-fecb-4c95-b487-94dcfc651311-logs\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.556339 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6f61a35-fecb-4c95-b487-94dcfc651311-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.557686 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6f61a35-fecb-4c95-b487-94dcfc651311-config-data\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.559689 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6f61a35-fecb-4c95-b487-94dcfc651311-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.571339 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6rgf\" (UniqueName: \"kubernetes.io/projected/a6f61a35-fecb-4c95-b487-94dcfc651311-kube-api-access-m6rgf\") pod \"nova-metadata-0\" (UID: \"a6f61a35-fecb-4c95-b487-94dcfc651311\") " pod="openstack/nova-metadata-0" Dec 17 08:43:03 crc kubenswrapper[4966]: I1217 08:43:03.746446 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 17 08:43:04 crc kubenswrapper[4966]: I1217 08:43:04.259364 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 17 08:43:04 crc kubenswrapper[4966]: W1217 08:43:04.262415 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6f61a35_fecb_4c95_b487_94dcfc651311.slice/crio-393a6839b1bfe04002ccdd94510f6b503b8a191ba9b97a502af0b8234445db85 WatchSource:0}: Error finding container 393a6839b1bfe04002ccdd94510f6b503b8a191ba9b97a502af0b8234445db85: Status 404 returned error can't find the container with id 393a6839b1bfe04002ccdd94510f6b503b8a191ba9b97a502af0b8234445db85 Dec 17 08:43:04 crc kubenswrapper[4966]: E1217 08:43:04.707681 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e is running failed: container process not found" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 17 08:43:04 crc kubenswrapper[4966]: E1217 08:43:04.708444 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e is running failed: container process not found" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 17 08:43:04 crc kubenswrapper[4966]: E1217 08:43:04.708996 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e is running failed: container process not found" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 17 08:43:04 crc kubenswrapper[4966]: E1217 08:43:04.709039 4966 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="03b5489a-56c8-441a-8fb6-421ac6275281" containerName="nova-scheduler-scheduler" Dec 17 08:43:04 crc kubenswrapper[4966]: I1217 08:43:04.745208 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:43:04 crc kubenswrapper[4966]: I1217 08:43:04.861252 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" path="/var/lib/kubelet/pods/d3b5f2dd-052a-4bb8-afaf-96b9ef596219/volumes" Dec 17 08:43:04 crc kubenswrapper[4966]: I1217 08:43:04.899581 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trjpt\" (UniqueName: \"kubernetes.io/projected/03b5489a-56c8-441a-8fb6-421ac6275281-kube-api-access-trjpt\") pod \"03b5489a-56c8-441a-8fb6-421ac6275281\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " Dec 17 08:43:04 crc kubenswrapper[4966]: I1217 08:43:04.899702 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-config-data\") pod \"03b5489a-56c8-441a-8fb6-421ac6275281\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " Dec 17 08:43:04 crc kubenswrapper[4966]: I1217 08:43:04.899743 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-combined-ca-bundle\") pod \"03b5489a-56c8-441a-8fb6-421ac6275281\" (UID: \"03b5489a-56c8-441a-8fb6-421ac6275281\") " Dec 17 08:43:04 crc kubenswrapper[4966]: I1217 08:43:04.912136 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03b5489a-56c8-441a-8fb6-421ac6275281-kube-api-access-trjpt" (OuterVolumeSpecName: "kube-api-access-trjpt") pod "03b5489a-56c8-441a-8fb6-421ac6275281" (UID: "03b5489a-56c8-441a-8fb6-421ac6275281"). InnerVolumeSpecName "kube-api-access-trjpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:43:04 crc kubenswrapper[4966]: I1217 08:43:04.956008 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-config-data" (OuterVolumeSpecName: "config-data") pod "03b5489a-56c8-441a-8fb6-421ac6275281" (UID: "03b5489a-56c8-441a-8fb6-421ac6275281"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:04 crc kubenswrapper[4966]: I1217 08:43:04.980140 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03b5489a-56c8-441a-8fb6-421ac6275281" (UID: "03b5489a-56c8-441a-8fb6-421ac6275281"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.003666 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trjpt\" (UniqueName: \"kubernetes.io/projected/03b5489a-56c8-441a-8fb6-421ac6275281-kube-api-access-trjpt\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.003701 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.003711 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03b5489a-56c8-441a-8fb6-421ac6275281-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.087097 4966 generic.go:334] "Generic (PLEG): container finished" podID="03b5489a-56c8-441a-8fb6-421ac6275281" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" exitCode=0 Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.087154 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"03b5489a-56c8-441a-8fb6-421ac6275281","Type":"ContainerDied","Data":"acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e"} Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.087166 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.087238 4966 scope.go:117] "RemoveContainer" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.087223 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"03b5489a-56c8-441a-8fb6-421ac6275281","Type":"ContainerDied","Data":"25e03a3af423b14658801279e13115dec81085d56fa0fd2e3b4fd1f91a6ccd22"} Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.089766 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a6f61a35-fecb-4c95-b487-94dcfc651311","Type":"ContainerStarted","Data":"77633ef20a369dd646def8e42b499795a7150a47120700c98126e0f9e7843d3e"} Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.089808 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a6f61a35-fecb-4c95-b487-94dcfc651311","Type":"ContainerStarted","Data":"393a6839b1bfe04002ccdd94510f6b503b8a191ba9b97a502af0b8234445db85"} Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.117492 4966 scope.go:117] "RemoveContainer" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.118764 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:43:05 crc kubenswrapper[4966]: E1217 08:43:05.119678 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e\": container with ID starting with acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e not found: ID does not exist" containerID="acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.119777 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e"} err="failed to get container status \"acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e\": rpc error: code = NotFound desc = could not find container \"acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e\": container with ID starting with acefc0a3d5781f62c07e020398c079314234f3262de25d32ce4640cf84a34a2e not found: ID does not exist" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.130048 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.140539 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:43:05 crc kubenswrapper[4966]: E1217 08:43:05.140992 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03b5489a-56c8-441a-8fb6-421ac6275281" containerName="nova-scheduler-scheduler" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.141010 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="03b5489a-56c8-441a-8fb6-421ac6275281" containerName="nova-scheduler-scheduler" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.141279 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="03b5489a-56c8-441a-8fb6-421ac6275281" containerName="nova-scheduler-scheduler" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.142181 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.144190 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.171924 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.308104 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a30b4f-0765-46df-900b-27d99cb78c50-config-data\") pod \"nova-scheduler-0\" (UID: \"07a30b4f-0765-46df-900b-27d99cb78c50\") " pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.308178 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg7kc\" (UniqueName: \"kubernetes.io/projected/07a30b4f-0765-46df-900b-27d99cb78c50-kube-api-access-xg7kc\") pod \"nova-scheduler-0\" (UID: \"07a30b4f-0765-46df-900b-27d99cb78c50\") " pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.308426 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a30b4f-0765-46df-900b-27d99cb78c50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"07a30b4f-0765-46df-900b-27d99cb78c50\") " pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.410173 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a30b4f-0765-46df-900b-27d99cb78c50-config-data\") pod \"nova-scheduler-0\" (UID: \"07a30b4f-0765-46df-900b-27d99cb78c50\") " pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.410263 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg7kc\" (UniqueName: \"kubernetes.io/projected/07a30b4f-0765-46df-900b-27d99cb78c50-kube-api-access-xg7kc\") pod \"nova-scheduler-0\" (UID: \"07a30b4f-0765-46df-900b-27d99cb78c50\") " pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.410359 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a30b4f-0765-46df-900b-27d99cb78c50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"07a30b4f-0765-46df-900b-27d99cb78c50\") " pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.422253 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07a30b4f-0765-46df-900b-27d99cb78c50-config-data\") pod \"nova-scheduler-0\" (UID: \"07a30b4f-0765-46df-900b-27d99cb78c50\") " pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.422278 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a30b4f-0765-46df-900b-27d99cb78c50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"07a30b4f-0765-46df-900b-27d99cb78c50\") " pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.428579 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg7kc\" (UniqueName: \"kubernetes.io/projected/07a30b4f-0765-46df-900b-27d99cb78c50-kube-api-access-xg7kc\") pod \"nova-scheduler-0\" (UID: \"07a30b4f-0765-46df-900b-27d99cb78c50\") " pod="openstack/nova-scheduler-0" Dec 17 08:43:05 crc kubenswrapper[4966]: I1217 08:43:05.484305 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.020422 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.128771 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a6f61a35-fecb-4c95-b487-94dcfc651311","Type":"ContainerStarted","Data":"c2c80ec6d3456477409576d3a977f4337fd19064a2bc4b767b283e138c5d32ef"} Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.134421 4966 generic.go:334] "Generic (PLEG): container finished" podID="036fc8b2-4f95-496c-b364-29de6b452030" containerID="3864b2899f39f6c2897d12bdbcc9d0e9465fb5e8a94aa91ec8a4b42a31468ef8" exitCode=0 Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.134634 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036fc8b2-4f95-496c-b364-29de6b452030","Type":"ContainerDied","Data":"3864b2899f39f6c2897d12bdbcc9d0e9465fb5e8a94aa91ec8a4b42a31468ef8"} Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.166064 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"07a30b4f-0765-46df-900b-27d99cb78c50","Type":"ContainerStarted","Data":"e0ec1463f6617cbfe95462e8d7936628bac56188f7db60da007560cba57d06b0"} Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.180172 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.180150006 podStartE2EDuration="3.180150006s" podCreationTimestamp="2025-12-17 08:43:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:43:06.156389354 +0000 UTC m=+1321.701459306" watchObservedRunningTime="2025-12-17 08:43:06.180150006 +0000 UTC m=+1321.725219948" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.258565 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.333194 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-internal-tls-certs\") pod \"036fc8b2-4f95-496c-b364-29de6b452030\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.333315 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-public-tls-certs\") pod \"036fc8b2-4f95-496c-b364-29de6b452030\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.333432 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-combined-ca-bundle\") pod \"036fc8b2-4f95-496c-b364-29de6b452030\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.333489 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036fc8b2-4f95-496c-b364-29de6b452030-logs\") pod \"036fc8b2-4f95-496c-b364-29de6b452030\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.333555 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-config-data\") pod \"036fc8b2-4f95-496c-b364-29de6b452030\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.333579 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zlch\" (UniqueName: \"kubernetes.io/projected/036fc8b2-4f95-496c-b364-29de6b452030-kube-api-access-4zlch\") pod \"036fc8b2-4f95-496c-b364-29de6b452030\" (UID: \"036fc8b2-4f95-496c-b364-29de6b452030\") " Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.337551 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/036fc8b2-4f95-496c-b364-29de6b452030-logs" (OuterVolumeSpecName: "logs") pod "036fc8b2-4f95-496c-b364-29de6b452030" (UID: "036fc8b2-4f95-496c-b364-29de6b452030"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.340714 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/036fc8b2-4f95-496c-b364-29de6b452030-kube-api-access-4zlch" (OuterVolumeSpecName: "kube-api-access-4zlch") pod "036fc8b2-4f95-496c-b364-29de6b452030" (UID: "036fc8b2-4f95-496c-b364-29de6b452030"). InnerVolumeSpecName "kube-api-access-4zlch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.400650 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-config-data" (OuterVolumeSpecName: "config-data") pod "036fc8b2-4f95-496c-b364-29de6b452030" (UID: "036fc8b2-4f95-496c-b364-29de6b452030"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.405274 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "036fc8b2-4f95-496c-b364-29de6b452030" (UID: "036fc8b2-4f95-496c-b364-29de6b452030"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.436098 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.436123 4966 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036fc8b2-4f95-496c-b364-29de6b452030-logs\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.436134 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.436145 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zlch\" (UniqueName: \"kubernetes.io/projected/036fc8b2-4f95-496c-b364-29de6b452030-kube-api-access-4zlch\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.467433 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "036fc8b2-4f95-496c-b364-29de6b452030" (UID: "036fc8b2-4f95-496c-b364-29de6b452030"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.470020 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "036fc8b2-4f95-496c-b364-29de6b452030" (UID: "036fc8b2-4f95-496c-b364-29de6b452030"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.538413 4966 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.538454 4966 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036fc8b2-4f95-496c-b364-29de6b452030-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:06 crc kubenswrapper[4966]: I1217 08:43:06.874334 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03b5489a-56c8-441a-8fb6-421ac6275281" path="/var/lib/kubelet/pods/03b5489a-56c8-441a-8fb6-421ac6275281/volumes" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.175488 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"07a30b4f-0765-46df-900b-27d99cb78c50","Type":"ContainerStarted","Data":"a26760df9a5fc71f30e4bfc2fea9bbcf5e2c36546efb19007050bb9745ce5436"} Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.178804 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.179381 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"036fc8b2-4f95-496c-b364-29de6b452030","Type":"ContainerDied","Data":"c8722be9b9ea15b0196bb8beb6aafdc0cf09036a3c2f5dd481a89978c025b0d0"} Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.179432 4966 scope.go:117] "RemoveContainer" containerID="3864b2899f39f6c2897d12bdbcc9d0e9465fb5e8a94aa91ec8a4b42a31468ef8" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.207945 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.207927648 podStartE2EDuration="2.207927648s" podCreationTimestamp="2025-12-17 08:43:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:43:07.20469109 +0000 UTC m=+1322.749761042" watchObservedRunningTime="2025-12-17 08:43:07.207927648 +0000 UTC m=+1322.752997590" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.216556 4966 scope.go:117] "RemoveContainer" containerID="d700a688723c1a8be0871b4f375e112861798ee6d37ce059556167d96721dd43" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.232995 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.243260 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.250785 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 17 08:43:07 crc kubenswrapper[4966]: E1217 08:43:07.251784 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-log" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.251804 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-log" Dec 17 08:43:07 crc kubenswrapper[4966]: E1217 08:43:07.251819 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-api" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.251826 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-api" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.252018 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-api" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.252034 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="036fc8b2-4f95-496c-b364-29de6b452030" containerName="nova-api-log" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.252960 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.258431 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.258624 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.258753 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.270170 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.356588 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62kgc\" (UniqueName: \"kubernetes.io/projected/071983f0-c849-4d55-a9e6-1951ed3e50b3-kube-api-access-62kgc\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.356651 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.356742 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/071983f0-c849-4d55-a9e6-1951ed3e50b3-logs\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.356777 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.356814 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-config-data\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.356946 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-public-tls-certs\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.458242 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/071983f0-c849-4d55-a9e6-1951ed3e50b3-logs\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.458551 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.458592 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-config-data\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.458664 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/071983f0-c849-4d55-a9e6-1951ed3e50b3-logs\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.458681 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-public-tls-certs\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.458857 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62kgc\" (UniqueName: \"kubernetes.io/projected/071983f0-c849-4d55-a9e6-1951ed3e50b3-kube-api-access-62kgc\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.458964 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.463887 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.464098 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-config-data\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.469457 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-public-tls-certs\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.471104 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/071983f0-c849-4d55-a9e6-1951ed3e50b3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.479446 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62kgc\" (UniqueName: \"kubernetes.io/projected/071983f0-c849-4d55-a9e6-1951ed3e50b3-kube-api-access-62kgc\") pod \"nova-api-0\" (UID: \"071983f0-c849-4d55-a9e6-1951ed3e50b3\") " pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.574584 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.722764 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 17 08:43:07 crc kubenswrapper[4966]: I1217 08:43:07.722787 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="d3b5f2dd-052a-4bb8-afaf-96b9ef596219" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": dial tcp 10.217.0.211:8775: i/o timeout" Dec 17 08:43:08 crc kubenswrapper[4966]: I1217 08:43:08.088567 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 17 08:43:08 crc kubenswrapper[4966]: I1217 08:43:08.196022 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"071983f0-c849-4d55-a9e6-1951ed3e50b3","Type":"ContainerStarted","Data":"fff2c39a2b4c5ff4109f27656369c2b851c0371c1975f421153ace53dee2e84d"} Dec 17 08:43:08 crc kubenswrapper[4966]: I1217 08:43:08.747050 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 17 08:43:08 crc kubenswrapper[4966]: I1217 08:43:08.747438 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 17 08:43:08 crc kubenswrapper[4966]: I1217 08:43:08.840134 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="036fc8b2-4f95-496c-b364-29de6b452030" path="/var/lib/kubelet/pods/036fc8b2-4f95-496c-b364-29de6b452030/volumes" Dec 17 08:43:09 crc kubenswrapper[4966]: I1217 08:43:09.206933 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"071983f0-c849-4d55-a9e6-1951ed3e50b3","Type":"ContainerStarted","Data":"c621488c04aa275471c375c9a6b948749077f837b8e7f26b7271bbb04c5c14f5"} Dec 17 08:43:09 crc kubenswrapper[4966]: I1217 08:43:09.207188 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"071983f0-c849-4d55-a9e6-1951ed3e50b3","Type":"ContainerStarted","Data":"82445534a3a798582ad266d8b94aa697eb1abc0628bbb15e51c9d488a24e4458"} Dec 17 08:43:09 crc kubenswrapper[4966]: I1217 08:43:09.229768 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.229748588 podStartE2EDuration="2.229748588s" podCreationTimestamp="2025-12-17 08:43:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:43:09.224948177 +0000 UTC m=+1324.770018119" watchObservedRunningTime="2025-12-17 08:43:09.229748588 +0000 UTC m=+1324.774818530" Dec 17 08:43:10 crc kubenswrapper[4966]: I1217 08:43:10.485238 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 17 08:43:13 crc kubenswrapper[4966]: I1217 08:43:13.747219 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 17 08:43:13 crc kubenswrapper[4966]: I1217 08:43:13.750366 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 17 08:43:14 crc kubenswrapper[4966]: I1217 08:43:14.760185 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a6f61a35-fecb-4c95-b487-94dcfc651311" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.222:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 08:43:14 crc kubenswrapper[4966]: I1217 08:43:14.764849 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a6f61a35-fecb-4c95-b487-94dcfc651311" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.222:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 08:43:15 crc kubenswrapper[4966]: I1217 08:43:15.484893 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 17 08:43:15 crc kubenswrapper[4966]: I1217 08:43:15.521013 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 17 08:43:16 crc kubenswrapper[4966]: I1217 08:43:16.325602 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 17 08:43:17 crc kubenswrapper[4966]: I1217 08:43:17.575636 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 17 08:43:17 crc kubenswrapper[4966]: I1217 08:43:17.576992 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 17 08:43:18 crc kubenswrapper[4966]: I1217 08:43:18.593299 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="071983f0-c849-4d55-a9e6-1951ed3e50b3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.224:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 08:43:18 crc kubenswrapper[4966]: I1217 08:43:18.594996 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="071983f0-c849-4d55-a9e6-1951ed3e50b3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.224:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 08:43:23 crc kubenswrapper[4966]: I1217 08:43:23.755163 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 17 08:43:23 crc kubenswrapper[4966]: I1217 08:43:23.763750 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 17 08:43:23 crc kubenswrapper[4966]: I1217 08:43:23.772258 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 17 08:43:24 crc kubenswrapper[4966]: I1217 08:43:24.390260 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 17 08:43:27 crc kubenswrapper[4966]: I1217 08:43:27.588213 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 17 08:43:27 crc kubenswrapper[4966]: I1217 08:43:27.590638 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 17 08:43:27 crc kubenswrapper[4966]: I1217 08:43:27.594426 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 17 08:43:27 crc kubenswrapper[4966]: I1217 08:43:27.598525 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 17 08:43:28 crc kubenswrapper[4966]: I1217 08:43:28.440280 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 17 08:43:28 crc kubenswrapper[4966]: I1217 08:43:28.447397 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 17 08:43:28 crc kubenswrapper[4966]: I1217 08:43:28.484987 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 17 08:43:38 crc kubenswrapper[4966]: I1217 08:43:38.183160 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 17 08:43:38 crc kubenswrapper[4966]: I1217 08:43:38.997421 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 17 08:43:43 crc kubenswrapper[4966]: I1217 08:43:43.044221 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" containerName="rabbitmq" containerID="cri-o://08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392" gracePeriod=604796 Dec 17 08:43:43 crc kubenswrapper[4966]: I1217 08:43:43.589378 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="1bd01548-3d59-4472-b8ef-a899b2cedee8" containerName="rabbitmq" containerID="cri-o://c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90" gracePeriod=604796 Dec 17 08:43:48 crc kubenswrapper[4966]: I1217 08:43:48.534044 4966 scope.go:117] "RemoveContainer" containerID="1ca1f40e30419dcae865d389041d39c16f770b817813aaaf6b8415cfc0e53fbf" Dec 17 08:43:48 crc kubenswrapper[4966]: I1217 08:43:48.568620 4966 scope.go:117] "RemoveContainer" containerID="0ab9548d146e4a17c461faa032ae051729fabb30139069121427c5f6dc8f338a" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.535710 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="1bd01548-3d59-4472-b8ef-a899b2cedee8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.105:5671: connect: connection refused" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.623952 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.694091 4966 generic.go:334] "Generic (PLEG): container finished" podID="6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" containerID="08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392" exitCode=0 Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.694413 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb","Type":"ContainerDied","Data":"08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392"} Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.694446 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb","Type":"ContainerDied","Data":"4f8932c4782c2bda7d35a0975dd831c49dafc9e0b31ab55f6bd32af612a1a7bc"} Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.694472 4966 scope.go:117] "RemoveContainer" containerID="08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.694622 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.753832 4966 scope.go:117] "RemoveContainer" containerID="40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.767657 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-tls\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.767709 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccbzf\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-kube-api-access-ccbzf\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.767736 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-plugins-conf\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.767841 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-pod-info\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.767888 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-server-conf\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.767914 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-erlang-cookie\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.767955 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.767977 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-confd\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.768007 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-plugins\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.768058 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-erlang-cookie-secret\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.768092 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-config-data\") pod \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\" (UID: \"6eeef4d2-f245-4de5-a0c5-fa9abe322dcb\") " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.768932 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.769408 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.771646 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.777953 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-kube-api-access-ccbzf" (OuterVolumeSpecName: "kube-api-access-ccbzf") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "kube-api-access-ccbzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.778490 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.778885 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.780990 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-pod-info" (OuterVolumeSpecName: "pod-info") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.782958 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.807218 4966 scope.go:117] "RemoveContainer" containerID="08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392" Dec 17 08:43:49 crc kubenswrapper[4966]: E1217 08:43:49.807674 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392\": container with ID starting with 08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392 not found: ID does not exist" containerID="08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.807705 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392"} err="failed to get container status \"08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392\": rpc error: code = NotFound desc = could not find container \"08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392\": container with ID starting with 08b5641e58d70b50254cc25713df692d856f013dfed4f99f04a81a097a42e392 not found: ID does not exist" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.807725 4966 scope.go:117] "RemoveContainer" containerID="40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a" Dec 17 08:43:49 crc kubenswrapper[4966]: E1217 08:43:49.808405 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a\": container with ID starting with 40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a not found: ID does not exist" containerID="40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.808427 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a"} err="failed to get container status \"40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a\": rpc error: code = NotFound desc = could not find container \"40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a\": container with ID starting with 40402fb5dfeb5e97903a09141f80015ece22f7883d965f6d7361a9f6b4cf7c8a not found: ID does not exist" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.809936 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-config-data" (OuterVolumeSpecName: "config-data") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.869729 4966 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.869945 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccbzf\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-kube-api-access-ccbzf\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.870066 4966 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.870146 4966 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-pod-info\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.870235 4966 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.870324 4966 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.870407 4966 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.870481 4966 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.870561 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.893741 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-server-conf" (OuterVolumeSpecName: "server-conf") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.922806 4966 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.973155 4966 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-server-conf\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.973415 4966 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:49 crc kubenswrapper[4966]: I1217 08:43:49.991596 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" (UID: "6eeef4d2-f245-4de5-a0c5-fa9abe322dcb"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.080508 4966 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.092928 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.102070 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.130266 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 17 08:43:50 crc kubenswrapper[4966]: E1217 08:43:50.131004 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" containerName="setup-container" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.131023 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" containerName="setup-container" Dec 17 08:43:50 crc kubenswrapper[4966]: E1217 08:43:50.131042 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" containerName="rabbitmq" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.131049 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" containerName="rabbitmq" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.131296 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" containerName="rabbitmq" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.132643 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.139161 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.139208 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.139409 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vj9tw" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.139773 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.143523 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.143639 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.143711 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.165249 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.234662 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283654 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283714 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283738 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283757 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283818 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283845 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283864 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283903 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283925 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283970 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9w7d\" (UniqueName: \"kubernetes.io/projected/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-kube-api-access-d9w7d\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.283998 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.329864 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-599c884575-zmns2"] Dec 17 08:43:50 crc kubenswrapper[4966]: E1217 08:43:50.330660 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bd01548-3d59-4472-b8ef-a899b2cedee8" containerName="setup-container" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.330682 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bd01548-3d59-4472-b8ef-a899b2cedee8" containerName="setup-container" Dec 17 08:43:50 crc kubenswrapper[4966]: E1217 08:43:50.330696 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bd01548-3d59-4472-b8ef-a899b2cedee8" containerName="rabbitmq" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.330707 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bd01548-3d59-4472-b8ef-a899b2cedee8" containerName="rabbitmq" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.331135 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bd01548-3d59-4472-b8ef-a899b2cedee8" containerName="rabbitmq" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.337942 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.340592 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.366514 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-599c884575-zmns2"] Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.387465 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.387758 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-confd\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.387859 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-plugins\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.387969 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-tls\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.388074 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bd01548-3d59-4472-b8ef-a899b2cedee8-pod-info\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.388187 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-config-data\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.388274 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-server-conf\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.388370 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rbpr\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-kube-api-access-9rbpr\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.388484 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bd01548-3d59-4472-b8ef-a899b2cedee8-erlang-cookie-secret\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.388561 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-plugins-conf\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.388639 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-erlang-cookie\") pod \"1bd01548-3d59-4472-b8ef-a899b2cedee8\" (UID: \"1bd01548-3d59-4472-b8ef-a899b2cedee8\") " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.389042 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.389190 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.389304 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.389411 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.389514 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.389712 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.389847 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.390003 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.390130 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.390220 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.390362 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9w7d\" (UniqueName: \"kubernetes.io/projected/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-kube-api-access-d9w7d\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.395168 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1bd01548-3d59-4472-b8ef-a899b2cedee8-pod-info" (OuterVolumeSpecName: "pod-info") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.395359 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.396205 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.396466 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.396665 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.401777 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.403068 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.403296 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.406247 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.412398 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.415741 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.417489 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-kube-api-access-9rbpr" (OuterVolumeSpecName: "kube-api-access-9rbpr") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "kube-api-access-9rbpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.419089 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.423601 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bd01548-3d59-4472-b8ef-a899b2cedee8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.427378 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.446536 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.461213 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9w7d\" (UniqueName: \"kubernetes.io/projected/c0dca84d-27d4-4e9a-a9d5-303031bb71a2-kube-api-access-d9w7d\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.462772 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.463199 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.473819 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-config-data" (OuterVolumeSpecName: "config-data") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.492416 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-swift-storage-0\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.492626 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-svc\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.492783 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-config\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.492857 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lf5b\" (UniqueName: \"kubernetes.io/projected/091cc17e-9418-45cf-ac3a-6a07c6d27966-kube-api-access-8lf5b\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.492999 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-sb\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493141 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-nb\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493246 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-openstack-edpm-ipam\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493375 4966 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493437 4966 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493492 4966 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1bd01548-3d59-4472-b8ef-a899b2cedee8-pod-info\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493545 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493606 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rbpr\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-kube-api-access-9rbpr\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493680 4966 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1bd01548-3d59-4472-b8ef-a899b2cedee8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493746 4966 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493800 4966 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.493911 4966 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.497280 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c0dca84d-27d4-4e9a-a9d5-303031bb71a2\") " pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.516146 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-server-conf" (OuterVolumeSpecName: "server-conf") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.525141 4966 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.590908 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "1bd01548-3d59-4472-b8ef-a899b2cedee8" (UID: "1bd01548-3d59-4472-b8ef-a899b2cedee8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.598705 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-nb\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.599744 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-nb\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.600263 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-openstack-edpm-ipam\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.600945 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-openstack-edpm-ipam\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.601210 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-svc\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.601818 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-svc\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.602079 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-swift-storage-0\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.602335 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-config\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.603143 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lf5b\" (UniqueName: \"kubernetes.io/projected/091cc17e-9418-45cf-ac3a-6a07c6d27966-kube-api-access-8lf5b\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.603085 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-config\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.602761 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-swift-storage-0\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.604068 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-sb\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.604774 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-sb\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.605068 4966 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.605169 4966 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1bd01548-3d59-4472-b8ef-a899b2cedee8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.605259 4966 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1bd01548-3d59-4472-b8ef-a899b2cedee8-server-conf\") on node \"crc\" DevicePath \"\"" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.620565 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lf5b\" (UniqueName: \"kubernetes.io/projected/091cc17e-9418-45cf-ac3a-6a07c6d27966-kube-api-access-8lf5b\") pod \"dnsmasq-dns-599c884575-zmns2\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.661728 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.718770 4966 generic.go:334] "Generic (PLEG): container finished" podID="1bd01548-3d59-4472-b8ef-a899b2cedee8" containerID="c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90" exitCode=0 Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.718816 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1bd01548-3d59-4472-b8ef-a899b2cedee8","Type":"ContainerDied","Data":"c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90"} Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.718861 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1bd01548-3d59-4472-b8ef-a899b2cedee8","Type":"ContainerDied","Data":"20b0f6f21e431555cb4d03aedbaf0f13fdd6ed875e549f3b4c4630c4cb5a49ba"} Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.718891 4966 scope.go:117] "RemoveContainer" containerID="c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.718903 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.773991 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.784101 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.787676 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.790832 4966 scope.go:117] "RemoveContainer" containerID="f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.826055 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.828293 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.846414 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.846818 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.846931 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.846814 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.847193 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ft6gd" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.847320 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.854164 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.881372 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bd01548-3d59-4472-b8ef-a899b2cedee8" path="/var/lib/kubelet/pods/1bd01548-3d59-4472-b8ef-a899b2cedee8/volumes" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.882763 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eeef4d2-f245-4de5-a0c5-fa9abe322dcb" path="/var/lib/kubelet/pods/6eeef4d2-f245-4de5-a0c5-fa9abe322dcb/volumes" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.903863 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919232 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919269 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919318 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4610130f-ad82-401c-a58b-5edd37f31d1e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919360 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919411 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4610130f-ad82-401c-a58b-5edd37f31d1e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919458 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4610130f-ad82-401c-a58b-5edd37f31d1e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919502 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4610130f-ad82-401c-a58b-5edd37f31d1e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919526 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rczlf\" (UniqueName: \"kubernetes.io/projected/4610130f-ad82-401c-a58b-5edd37f31d1e-kube-api-access-rczlf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919561 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4610130f-ad82-401c-a58b-5edd37f31d1e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919589 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.919613 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.943407 4966 scope.go:117] "RemoveContainer" containerID="c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90" Dec 17 08:43:50 crc kubenswrapper[4966]: E1217 08:43:50.943785 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90\": container with ID starting with c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90 not found: ID does not exist" containerID="c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.943816 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90"} err="failed to get container status \"c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90\": rpc error: code = NotFound desc = could not find container \"c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90\": container with ID starting with c4d4e9c8e3a03e4c1da60cbd72c8f444a3c568f5bf5d76cac5512a217ebe7f90 not found: ID does not exist" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.943835 4966 scope.go:117] "RemoveContainer" containerID="f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd" Dec 17 08:43:50 crc kubenswrapper[4966]: E1217 08:43:50.944523 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd\": container with ID starting with f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd not found: ID does not exist" containerID="f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd" Dec 17 08:43:50 crc kubenswrapper[4966]: I1217 08:43:50.944573 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd"} err="failed to get container status \"f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd\": rpc error: code = NotFound desc = could not find container \"f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd\": container with ID starting with f7074f159f475d932d8dcd33e11a70128b36f4e59adc46324b7d99e0f6373bfd not found: ID does not exist" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028321 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczlf\" (UniqueName: \"kubernetes.io/projected/4610130f-ad82-401c-a58b-5edd37f31d1e-kube-api-access-rczlf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028418 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4610130f-ad82-401c-a58b-5edd37f31d1e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028498 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028543 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028632 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028649 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028710 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4610130f-ad82-401c-a58b-5edd37f31d1e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028767 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028791 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4610130f-ad82-401c-a58b-5edd37f31d1e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028852 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4610130f-ad82-401c-a58b-5edd37f31d1e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.028924 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4610130f-ad82-401c-a58b-5edd37f31d1e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.030735 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.030971 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4610130f-ad82-401c-a58b-5edd37f31d1e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.031688 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4610130f-ad82-401c-a58b-5edd37f31d1e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.032778 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.033854 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.036359 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4610130f-ad82-401c-a58b-5edd37f31d1e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.041155 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4610130f-ad82-401c-a58b-5edd37f31d1e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.060228 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczlf\" (UniqueName: \"kubernetes.io/projected/4610130f-ad82-401c-a58b-5edd37f31d1e-kube-api-access-rczlf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.068315 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.069413 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4610130f-ad82-401c-a58b-5edd37f31d1e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.073744 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4610130f-ad82-401c-a58b-5edd37f31d1e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.107762 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"4610130f-ad82-401c-a58b-5edd37f31d1e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.229094 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-599c884575-zmns2"] Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.346432 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.467206 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.730156 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c0dca84d-27d4-4e9a-a9d5-303031bb71a2","Type":"ContainerStarted","Data":"94df0d48e844dd6b912a5d6b114895da87ed27a94970e588ded54deb191c2cdd"} Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.735426 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-599c884575-zmns2" event={"ID":"091cc17e-9418-45cf-ac3a-6a07c6d27966","Type":"ContainerStarted","Data":"c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713"} Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.735479 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-599c884575-zmns2" event={"ID":"091cc17e-9418-45cf-ac3a-6a07c6d27966","Type":"ContainerStarted","Data":"e30a04f70acbadaba7e08e31e8ce580b3090c4d61e9f8737f6ff6f2d9ad1a30b"} Dec 17 08:43:51 crc kubenswrapper[4966]: I1217 08:43:51.854867 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 17 08:43:51 crc kubenswrapper[4966]: W1217 08:43:51.879917 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4610130f_ad82_401c_a58b_5edd37f31d1e.slice/crio-c82fa4a7ed170cc10c9263a1a64a01dcd98078a223a3fa82ae0dfef6ae4b6f75 WatchSource:0}: Error finding container c82fa4a7ed170cc10c9263a1a64a01dcd98078a223a3fa82ae0dfef6ae4b6f75: Status 404 returned error can't find the container with id c82fa4a7ed170cc10c9263a1a64a01dcd98078a223a3fa82ae0dfef6ae4b6f75 Dec 17 08:43:52 crc kubenswrapper[4966]: I1217 08:43:52.745080 4966 generic.go:334] "Generic (PLEG): container finished" podID="091cc17e-9418-45cf-ac3a-6a07c6d27966" containerID="c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713" exitCode=0 Dec 17 08:43:52 crc kubenswrapper[4966]: I1217 08:43:52.745138 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-599c884575-zmns2" event={"ID":"091cc17e-9418-45cf-ac3a-6a07c6d27966","Type":"ContainerDied","Data":"c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713"} Dec 17 08:43:52 crc kubenswrapper[4966]: I1217 08:43:52.747207 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4610130f-ad82-401c-a58b-5edd37f31d1e","Type":"ContainerStarted","Data":"c82fa4a7ed170cc10c9263a1a64a01dcd98078a223a3fa82ae0dfef6ae4b6f75"} Dec 17 08:43:53 crc kubenswrapper[4966]: I1217 08:43:53.761299 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4610130f-ad82-401c-a58b-5edd37f31d1e","Type":"ContainerStarted","Data":"3c650461180b501610388bb553a4ba092a012860551ca008a0ed863271c76af8"} Dec 17 08:43:53 crc kubenswrapper[4966]: I1217 08:43:53.765133 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-599c884575-zmns2" event={"ID":"091cc17e-9418-45cf-ac3a-6a07c6d27966","Type":"ContainerStarted","Data":"2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763"} Dec 17 08:43:53 crc kubenswrapper[4966]: I1217 08:43:53.765203 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:43:53 crc kubenswrapper[4966]: I1217 08:43:53.766541 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c0dca84d-27d4-4e9a-a9d5-303031bb71a2","Type":"ContainerStarted","Data":"ea9e831da9c59c7b1ffec4c8eab0c66e3c5dd9c686c449ed7a6285d4e894705f"} Dec 17 08:43:53 crc kubenswrapper[4966]: I1217 08:43:53.855890 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-599c884575-zmns2" podStartSLOduration=3.855859907 podStartE2EDuration="3.855859907s" podCreationTimestamp="2025-12-17 08:43:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:43:53.848825234 +0000 UTC m=+1369.393895176" watchObservedRunningTime="2025-12-17 08:43:53.855859907 +0000 UTC m=+1369.400929849" Dec 17 08:44:00 crc kubenswrapper[4966]: I1217 08:44:00.663442 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:44:00 crc kubenswrapper[4966]: I1217 08:44:00.742921 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d65fcb98c-nd4qw"] Dec 17 08:44:00 crc kubenswrapper[4966]: I1217 08:44:00.746704 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" podUID="b64018c5-5835-4e79-908f-b2f96d279dcf" containerName="dnsmasq-dns" containerID="cri-o://762e174390ddcedf69063400ad72b0960633ed056a04985dbeaf65b3376401b8" gracePeriod=10 Dec 17 08:44:00 crc kubenswrapper[4966]: I1217 08:44:00.894477 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" event={"ID":"b64018c5-5835-4e79-908f-b2f96d279dcf","Type":"ContainerDied","Data":"762e174390ddcedf69063400ad72b0960633ed056a04985dbeaf65b3376401b8"} Dec 17 08:44:00 crc kubenswrapper[4966]: I1217 08:44:00.894808 4966 generic.go:334] "Generic (PLEG): container finished" podID="b64018c5-5835-4e79-908f-b2f96d279dcf" containerID="762e174390ddcedf69063400ad72b0960633ed056a04985dbeaf65b3376401b8" exitCode=0 Dec 17 08:44:00 crc kubenswrapper[4966]: I1217 08:44:00.934094 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5999df4fc9-v4kbm"] Dec 17 08:44:00 crc kubenswrapper[4966]: I1217 08:44:00.936613 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:00 crc kubenswrapper[4966]: I1217 08:44:00.950129 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5999df4fc9-v4kbm"] Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.072205 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-config\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.072268 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-dns-swift-storage-0\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.072292 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-dns-svc\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.072330 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcpn6\" (UniqueName: \"kubernetes.io/projected/453bafa1-1048-40ee-a4a0-fc888cc20f44-kube-api-access-gcpn6\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.072388 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-ovsdbserver-nb\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.072402 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-ovsdbserver-sb\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.072430 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-openstack-edpm-ipam\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.174294 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-ovsdbserver-nb\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.174336 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-ovsdbserver-sb\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.175206 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-ovsdbserver-nb\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.175223 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-openstack-edpm-ipam\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.175208 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-ovsdbserver-sb\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.175292 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-openstack-edpm-ipam\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.175321 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-config\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.175376 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-dns-swift-storage-0\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.175395 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-dns-svc\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.175445 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcpn6\" (UniqueName: \"kubernetes.io/projected/453bafa1-1048-40ee-a4a0-fc888cc20f44-kube-api-access-gcpn6\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.176130 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-config\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.176567 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-dns-svc\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.176642 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/453bafa1-1048-40ee-a4a0-fc888cc20f44-dns-swift-storage-0\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.195801 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcpn6\" (UniqueName: \"kubernetes.io/projected/453bafa1-1048-40ee-a4a0-fc888cc20f44-kube-api-access-gcpn6\") pod \"dnsmasq-dns-5999df4fc9-v4kbm\" (UID: \"453bafa1-1048-40ee-a4a0-fc888cc20f44\") " pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.281729 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.418577 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.586679 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxv4b\" (UniqueName: \"kubernetes.io/projected/b64018c5-5835-4e79-908f-b2f96d279dcf-kube-api-access-wxv4b\") pod \"b64018c5-5835-4e79-908f-b2f96d279dcf\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.586758 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-swift-storage-0\") pod \"b64018c5-5835-4e79-908f-b2f96d279dcf\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.586851 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-nb\") pod \"b64018c5-5835-4e79-908f-b2f96d279dcf\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.586918 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-config\") pod \"b64018c5-5835-4e79-908f-b2f96d279dcf\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.586958 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-sb\") pod \"b64018c5-5835-4e79-908f-b2f96d279dcf\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.586991 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-svc\") pod \"b64018c5-5835-4e79-908f-b2f96d279dcf\" (UID: \"b64018c5-5835-4e79-908f-b2f96d279dcf\") " Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.591524 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b64018c5-5835-4e79-908f-b2f96d279dcf-kube-api-access-wxv4b" (OuterVolumeSpecName: "kube-api-access-wxv4b") pod "b64018c5-5835-4e79-908f-b2f96d279dcf" (UID: "b64018c5-5835-4e79-908f-b2f96d279dcf"). InnerVolumeSpecName "kube-api-access-wxv4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.647042 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b64018c5-5835-4e79-908f-b2f96d279dcf" (UID: "b64018c5-5835-4e79-908f-b2f96d279dcf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.664749 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b64018c5-5835-4e79-908f-b2f96d279dcf" (UID: "b64018c5-5835-4e79-908f-b2f96d279dcf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.667405 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b64018c5-5835-4e79-908f-b2f96d279dcf" (UID: "b64018c5-5835-4e79-908f-b2f96d279dcf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.677075 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-config" (OuterVolumeSpecName: "config") pod "b64018c5-5835-4e79-908f-b2f96d279dcf" (UID: "b64018c5-5835-4e79-908f-b2f96d279dcf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.678000 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b64018c5-5835-4e79-908f-b2f96d279dcf" (UID: "b64018c5-5835-4e79-908f-b2f96d279dcf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.689025 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.689064 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.689077 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.689094 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.689115 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxv4b\" (UniqueName: \"kubernetes.io/projected/b64018c5-5835-4e79-908f-b2f96d279dcf-kube-api-access-wxv4b\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.689128 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b64018c5-5835-4e79-908f-b2f96d279dcf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.791925 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5999df4fc9-v4kbm"] Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.908201 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" event={"ID":"b64018c5-5835-4e79-908f-b2f96d279dcf","Type":"ContainerDied","Data":"474d02e5919afcb5ed019add438df7306c2fcb8fceca9800115b204a79112a00"} Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.908282 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.908512 4966 scope.go:117] "RemoveContainer" containerID="762e174390ddcedf69063400ad72b0960633ed056a04985dbeaf65b3376401b8" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.914106 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" event={"ID":"453bafa1-1048-40ee-a4a0-fc888cc20f44","Type":"ContainerStarted","Data":"1127be2cc274dceccc1fd776637f7ebc5756ecb8dea5dd2d2e0f7aad8ce53b7c"} Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.962059 4966 scope.go:117] "RemoveContainer" containerID="0ea1617d5ee58d3f84abe10a555338f954881a61ba1214d724a9325aa57c958e" Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.982277 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d65fcb98c-nd4qw"] Dec 17 08:44:01 crc kubenswrapper[4966]: I1217 08:44:01.989482 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d65fcb98c-nd4qw"] Dec 17 08:44:02 crc kubenswrapper[4966]: I1217 08:44:02.841226 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b64018c5-5835-4e79-908f-b2f96d279dcf" path="/var/lib/kubelet/pods/b64018c5-5835-4e79-908f-b2f96d279dcf/volumes" Dec 17 08:44:02 crc kubenswrapper[4966]: I1217 08:44:02.925618 4966 generic.go:334] "Generic (PLEG): container finished" podID="453bafa1-1048-40ee-a4a0-fc888cc20f44" containerID="b992b014e05c50c478cf40d60ab7dda202c3498ff9caaa02c69cfd362bb4fec2" exitCode=0 Dec 17 08:44:02 crc kubenswrapper[4966]: I1217 08:44:02.925655 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" event={"ID":"453bafa1-1048-40ee-a4a0-fc888cc20f44","Type":"ContainerDied","Data":"b992b014e05c50c478cf40d60ab7dda202c3498ff9caaa02c69cfd362bb4fec2"} Dec 17 08:44:03 crc kubenswrapper[4966]: I1217 08:44:03.943799 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" event={"ID":"453bafa1-1048-40ee-a4a0-fc888cc20f44","Type":"ContainerStarted","Data":"8dea041b4cd3102009f42a6868e2ceb3303940519652cec2dbe044a697f321dd"} Dec 17 08:44:03 crc kubenswrapper[4966]: I1217 08:44:03.944210 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:03 crc kubenswrapper[4966]: I1217 08:44:03.978129 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" podStartSLOduration=3.978108874 podStartE2EDuration="3.978108874s" podCreationTimestamp="2025-12-17 08:44:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:44:03.965098289 +0000 UTC m=+1379.510168251" watchObservedRunningTime="2025-12-17 08:44:03.978108874 +0000 UTC m=+1379.523178836" Dec 17 08:44:06 crc kubenswrapper[4966]: I1217 08:44:06.090128 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5d65fcb98c-nd4qw" podUID="b64018c5-5835-4e79-908f-b2f96d279dcf" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.216:5353: i/o timeout" Dec 17 08:44:11 crc kubenswrapper[4966]: I1217 08:44:11.285297 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5999df4fc9-v4kbm" Dec 17 08:44:11 crc kubenswrapper[4966]: I1217 08:44:11.410001 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-599c884575-zmns2"] Dec 17 08:44:11 crc kubenswrapper[4966]: I1217 08:44:11.410272 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-599c884575-zmns2" podUID="091cc17e-9418-45cf-ac3a-6a07c6d27966" containerName="dnsmasq-dns" containerID="cri-o://2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763" gracePeriod=10 Dec 17 08:44:11 crc kubenswrapper[4966]: I1217 08:44:11.919233 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.007532 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-openstack-edpm-ipam\") pod \"091cc17e-9418-45cf-ac3a-6a07c6d27966\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.007615 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-nb\") pod \"091cc17e-9418-45cf-ac3a-6a07c6d27966\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.007668 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lf5b\" (UniqueName: \"kubernetes.io/projected/091cc17e-9418-45cf-ac3a-6a07c6d27966-kube-api-access-8lf5b\") pod \"091cc17e-9418-45cf-ac3a-6a07c6d27966\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.007692 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-config\") pod \"091cc17e-9418-45cf-ac3a-6a07c6d27966\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.007706 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-sb\") pod \"091cc17e-9418-45cf-ac3a-6a07c6d27966\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.007836 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-svc\") pod \"091cc17e-9418-45cf-ac3a-6a07c6d27966\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.007891 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-swift-storage-0\") pod \"091cc17e-9418-45cf-ac3a-6a07c6d27966\" (UID: \"091cc17e-9418-45cf-ac3a-6a07c6d27966\") " Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.017643 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091cc17e-9418-45cf-ac3a-6a07c6d27966-kube-api-access-8lf5b" (OuterVolumeSpecName: "kube-api-access-8lf5b") pod "091cc17e-9418-45cf-ac3a-6a07c6d27966" (UID: "091cc17e-9418-45cf-ac3a-6a07c6d27966"). InnerVolumeSpecName "kube-api-access-8lf5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.052078 4966 generic.go:334] "Generic (PLEG): container finished" podID="091cc17e-9418-45cf-ac3a-6a07c6d27966" containerID="2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763" exitCode=0 Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.052133 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-599c884575-zmns2" event={"ID":"091cc17e-9418-45cf-ac3a-6a07c6d27966","Type":"ContainerDied","Data":"2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763"} Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.052150 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-599c884575-zmns2" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.052172 4966 scope.go:117] "RemoveContainer" containerID="2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.052160 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-599c884575-zmns2" event={"ID":"091cc17e-9418-45cf-ac3a-6a07c6d27966","Type":"ContainerDied","Data":"e30a04f70acbadaba7e08e31e8ce580b3090c4d61e9f8737f6ff6f2d9ad1a30b"} Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.066722 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "091cc17e-9418-45cf-ac3a-6a07c6d27966" (UID: "091cc17e-9418-45cf-ac3a-6a07c6d27966"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.075787 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-config" (OuterVolumeSpecName: "config") pod "091cc17e-9418-45cf-ac3a-6a07c6d27966" (UID: "091cc17e-9418-45cf-ac3a-6a07c6d27966"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.076723 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "091cc17e-9418-45cf-ac3a-6a07c6d27966" (UID: "091cc17e-9418-45cf-ac3a-6a07c6d27966"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.089220 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "091cc17e-9418-45cf-ac3a-6a07c6d27966" (UID: "091cc17e-9418-45cf-ac3a-6a07c6d27966"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.095521 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "091cc17e-9418-45cf-ac3a-6a07c6d27966" (UID: "091cc17e-9418-45cf-ac3a-6a07c6d27966"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.106476 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "091cc17e-9418-45cf-ac3a-6a07c6d27966" (UID: "091cc17e-9418-45cf-ac3a-6a07c6d27966"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.110984 4966 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.111018 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.111028 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lf5b\" (UniqueName: \"kubernetes.io/projected/091cc17e-9418-45cf-ac3a-6a07c6d27966-kube-api-access-8lf5b\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.111042 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-config\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.111050 4966 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.111060 4966 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.111068 4966 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/091cc17e-9418-45cf-ac3a-6a07c6d27966-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.129150 4966 scope.go:117] "RemoveContainer" containerID="c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.151655 4966 scope.go:117] "RemoveContainer" containerID="2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763" Dec 17 08:44:12 crc kubenswrapper[4966]: E1217 08:44:12.152095 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763\": container with ID starting with 2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763 not found: ID does not exist" containerID="2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.152136 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763"} err="failed to get container status \"2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763\": rpc error: code = NotFound desc = could not find container \"2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763\": container with ID starting with 2040c1896610f36e3df09e455c90a2b7af475e2bd9b205fce4d57c924c6da763 not found: ID does not exist" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.152162 4966 scope.go:117] "RemoveContainer" containerID="c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713" Dec 17 08:44:12 crc kubenswrapper[4966]: E1217 08:44:12.157206 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713\": container with ID starting with c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713 not found: ID does not exist" containerID="c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.157253 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713"} err="failed to get container status \"c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713\": rpc error: code = NotFound desc = could not find container \"c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713\": container with ID starting with c6f044fddcbf450243df436d68e0f88ae60e7a0a40991136d3acd22602982713 not found: ID does not exist" Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.399802 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-599c884575-zmns2"] Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.411351 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-599c884575-zmns2"] Dec 17 08:44:12 crc kubenswrapper[4966]: I1217 08:44:12.844206 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091cc17e-9418-45cf-ac3a-6a07c6d27966" path="/var/lib/kubelet/pods/091cc17e-9418-45cf-ac3a-6a07c6d27966/volumes" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.752518 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg"] Dec 17 08:44:25 crc kubenswrapper[4966]: E1217 08:44:25.753695 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64018c5-5835-4e79-908f-b2f96d279dcf" containerName="init" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.753710 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64018c5-5835-4e79-908f-b2f96d279dcf" containerName="init" Dec 17 08:44:25 crc kubenswrapper[4966]: E1217 08:44:25.753723 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b64018c5-5835-4e79-908f-b2f96d279dcf" containerName="dnsmasq-dns" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.753730 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b64018c5-5835-4e79-908f-b2f96d279dcf" containerName="dnsmasq-dns" Dec 17 08:44:25 crc kubenswrapper[4966]: E1217 08:44:25.753755 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091cc17e-9418-45cf-ac3a-6a07c6d27966" containerName="dnsmasq-dns" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.753762 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="091cc17e-9418-45cf-ac3a-6a07c6d27966" containerName="dnsmasq-dns" Dec 17 08:44:25 crc kubenswrapper[4966]: E1217 08:44:25.753980 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091cc17e-9418-45cf-ac3a-6a07c6d27966" containerName="init" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.753991 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="091cc17e-9418-45cf-ac3a-6a07c6d27966" containerName="init" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.754293 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="091cc17e-9418-45cf-ac3a-6a07c6d27966" containerName="dnsmasq-dns" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.754316 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b64018c5-5835-4e79-908f-b2f96d279dcf" containerName="dnsmasq-dns" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.755014 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.757895 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.758369 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.759994 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.760087 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.785524 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg"] Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.903948 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.904283 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.904394 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:25 crc kubenswrapper[4966]: I1217 08:44:25.904436 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbjzs\" (UniqueName: \"kubernetes.io/projected/8f32b388-d78c-42b8-9ac2-ba89910ca248-kube-api-access-jbjzs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.006823 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.007481 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.007736 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.007906 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbjzs\" (UniqueName: \"kubernetes.io/projected/8f32b388-d78c-42b8-9ac2-ba89910ca248-kube-api-access-jbjzs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.014656 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.016908 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.017934 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.027025 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbjzs\" (UniqueName: \"kubernetes.io/projected/8f32b388-d78c-42b8-9ac2-ba89910ca248-kube-api-access-jbjzs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.080227 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.216683 4966 generic.go:334] "Generic (PLEG): container finished" podID="4610130f-ad82-401c-a58b-5edd37f31d1e" containerID="3c650461180b501610388bb553a4ba092a012860551ca008a0ed863271c76af8" exitCode=0 Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.217038 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4610130f-ad82-401c-a58b-5edd37f31d1e","Type":"ContainerDied","Data":"3c650461180b501610388bb553a4ba092a012860551ca008a0ed863271c76af8"} Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.231765 4966 generic.go:334] "Generic (PLEG): container finished" podID="c0dca84d-27d4-4e9a-a9d5-303031bb71a2" containerID="ea9e831da9c59c7b1ffec4c8eab0c66e3c5dd9c686c449ed7a6285d4e894705f" exitCode=0 Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.231809 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c0dca84d-27d4-4e9a-a9d5-303031bb71a2","Type":"ContainerDied","Data":"ea9e831da9c59c7b1ffec4c8eab0c66e3c5dd9c686c449ed7a6285d4e894705f"} Dec 17 08:44:26 crc kubenswrapper[4966]: W1217 08:44:26.685433 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f32b388_d78c_42b8_9ac2_ba89910ca248.slice/crio-2cca30c0ee3090878148ec0d77da3e0b58672f5273a91a5979b1a88b2187a399 WatchSource:0}: Error finding container 2cca30c0ee3090878148ec0d77da3e0b58672f5273a91a5979b1a88b2187a399: Status 404 returned error can't find the container with id 2cca30c0ee3090878148ec0d77da3e0b58672f5273a91a5979b1a88b2187a399 Dec 17 08:44:26 crc kubenswrapper[4966]: I1217 08:44:26.693042 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg"] Dec 17 08:44:27 crc kubenswrapper[4966]: I1217 08:44:27.241932 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" event={"ID":"8f32b388-d78c-42b8-9ac2-ba89910ca248","Type":"ContainerStarted","Data":"2cca30c0ee3090878148ec0d77da3e0b58672f5273a91a5979b1a88b2187a399"} Dec 17 08:44:27 crc kubenswrapper[4966]: I1217 08:44:27.244745 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4610130f-ad82-401c-a58b-5edd37f31d1e","Type":"ContainerStarted","Data":"6bcd0f44a169e4cf8206628361c71a25514ddf648f0ebdffa5d0586f9b19bedd"} Dec 17 08:44:27 crc kubenswrapper[4966]: I1217 08:44:27.244988 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:44:27 crc kubenswrapper[4966]: I1217 08:44:27.247387 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c0dca84d-27d4-4e9a-a9d5-303031bb71a2","Type":"ContainerStarted","Data":"43326e422d06f937ffdbafeff926d9f2a5fb5cbe338e401ae499b36dbc9191f6"} Dec 17 08:44:27 crc kubenswrapper[4966]: I1217 08:44:27.247590 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 17 08:44:27 crc kubenswrapper[4966]: I1217 08:44:27.269910 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.269892555 podStartE2EDuration="37.269892555s" podCreationTimestamp="2025-12-17 08:43:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:44:27.265883246 +0000 UTC m=+1402.810953198" watchObservedRunningTime="2025-12-17 08:44:27.269892555 +0000 UTC m=+1402.814962507" Dec 17 08:44:27 crc kubenswrapper[4966]: I1217 08:44:27.296593 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.296575335 podStartE2EDuration="37.296575335s" podCreationTimestamp="2025-12-17 08:43:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 08:44:27.289400458 +0000 UTC m=+1402.834470400" watchObservedRunningTime="2025-12-17 08:44:27.296575335 +0000 UTC m=+1402.841645277" Dec 17 08:44:38 crc kubenswrapper[4966]: I1217 08:44:38.382949 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" event={"ID":"8f32b388-d78c-42b8-9ac2-ba89910ca248","Type":"ContainerStarted","Data":"286489e3a8e294c6ca1abfac16ccb229145d07a88e6a4e71f2a54cb26ba9731b"} Dec 17 08:44:38 crc kubenswrapper[4966]: I1217 08:44:38.407146 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" podStartSLOduration=2.536429943 podStartE2EDuration="13.407129876s" podCreationTimestamp="2025-12-17 08:44:25 +0000 UTC" firstStartedPulling="2025-12-17 08:44:26.68820819 +0000 UTC m=+1402.233278132" lastFinishedPulling="2025-12-17 08:44:37.558908133 +0000 UTC m=+1413.103978065" observedRunningTime="2025-12-17 08:44:38.399532737 +0000 UTC m=+1413.944602679" watchObservedRunningTime="2025-12-17 08:44:38.407129876 +0000 UTC m=+1413.952199818" Dec 17 08:44:40 crc kubenswrapper[4966]: I1217 08:44:40.787084 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 17 08:44:41 crc kubenswrapper[4966]: I1217 08:44:41.350049 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 17 08:44:46 crc kubenswrapper[4966]: I1217 08:44:46.807614 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:44:46 crc kubenswrapper[4966]: I1217 08:44:46.808295 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:44:48 crc kubenswrapper[4966]: I1217 08:44:48.790105 4966 scope.go:117] "RemoveContainer" containerID="cad9a1835e206a3215c09f9b7051c5339008db162d3ffc20a032223d28376b8a" Dec 17 08:44:48 crc kubenswrapper[4966]: I1217 08:44:48.821703 4966 scope.go:117] "RemoveContainer" containerID="8896caf4d77c1bdce82efba5b0a7f386f59c73c52fb327d369cc3355889b6597" Dec 17 08:44:59 crc kubenswrapper[4966]: I1217 08:44:59.637819 4966 generic.go:334] "Generic (PLEG): container finished" podID="8f32b388-d78c-42b8-9ac2-ba89910ca248" containerID="286489e3a8e294c6ca1abfac16ccb229145d07a88e6a4e71f2a54cb26ba9731b" exitCode=0 Dec 17 08:44:59 crc kubenswrapper[4966]: I1217 08:44:59.637901 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" event={"ID":"8f32b388-d78c-42b8-9ac2-ba89910ca248","Type":"ContainerDied","Data":"286489e3a8e294c6ca1abfac16ccb229145d07a88e6a4e71f2a54cb26ba9731b"} Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.159845 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8"] Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.162046 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.164863 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.165101 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.172648 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8"] Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.326831 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-secret-volume\") pod \"collect-profiles-29432685-7kjd8\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.326957 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-config-volume\") pod \"collect-profiles-29432685-7kjd8\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.327126 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qct7\" (UniqueName: \"kubernetes.io/projected/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-kube-api-access-2qct7\") pod \"collect-profiles-29432685-7kjd8\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.429178 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qct7\" (UniqueName: \"kubernetes.io/projected/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-kube-api-access-2qct7\") pod \"collect-profiles-29432685-7kjd8\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.429343 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-secret-volume\") pod \"collect-profiles-29432685-7kjd8\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.429460 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-config-volume\") pod \"collect-profiles-29432685-7kjd8\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.430659 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-config-volume\") pod \"collect-profiles-29432685-7kjd8\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.436700 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-secret-volume\") pod \"collect-profiles-29432685-7kjd8\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.460114 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qct7\" (UniqueName: \"kubernetes.io/projected/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-kube-api-access-2qct7\") pod \"collect-profiles-29432685-7kjd8\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:00 crc kubenswrapper[4966]: I1217 08:45:00.487491 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:01 crc kubenswrapper[4966]: W1217 08:45:01.029094 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b1a4d2c_b220_4bdb_8d2d_239b4e9f28f1.slice/crio-42c9e9889c083fb26cba43e9a088d4d46cb6bf87ff546f5437e8fb25e1442fda WatchSource:0}: Error finding container 42c9e9889c083fb26cba43e9a088d4d46cb6bf87ff546f5437e8fb25e1442fda: Status 404 returned error can't find the container with id 42c9e9889c083fb26cba43e9a088d4d46cb6bf87ff546f5437e8fb25e1442fda Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.030358 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8"] Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.084570 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.268431 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-repo-setup-combined-ca-bundle\") pod \"8f32b388-d78c-42b8-9ac2-ba89910ca248\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.268788 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-inventory\") pod \"8f32b388-d78c-42b8-9ac2-ba89910ca248\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.268999 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-ssh-key\") pod \"8f32b388-d78c-42b8-9ac2-ba89910ca248\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.269530 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbjzs\" (UniqueName: \"kubernetes.io/projected/8f32b388-d78c-42b8-9ac2-ba89910ca248-kube-api-access-jbjzs\") pod \"8f32b388-d78c-42b8-9ac2-ba89910ca248\" (UID: \"8f32b388-d78c-42b8-9ac2-ba89910ca248\") " Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.274636 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8f32b388-d78c-42b8-9ac2-ba89910ca248" (UID: "8f32b388-d78c-42b8-9ac2-ba89910ca248"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.275016 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f32b388-d78c-42b8-9ac2-ba89910ca248-kube-api-access-jbjzs" (OuterVolumeSpecName: "kube-api-access-jbjzs") pod "8f32b388-d78c-42b8-9ac2-ba89910ca248" (UID: "8f32b388-d78c-42b8-9ac2-ba89910ca248"). InnerVolumeSpecName "kube-api-access-jbjzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.301121 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-inventory" (OuterVolumeSpecName: "inventory") pod "8f32b388-d78c-42b8-9ac2-ba89910ca248" (UID: "8f32b388-d78c-42b8-9ac2-ba89910ca248"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.318061 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8f32b388-d78c-42b8-9ac2-ba89910ca248" (UID: "8f32b388-d78c-42b8-9ac2-ba89910ca248"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.372444 4966 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.372721 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.372938 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f32b388-d78c-42b8-9ac2-ba89910ca248-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.373069 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbjzs\" (UniqueName: \"kubernetes.io/projected/8f32b388-d78c-42b8-9ac2-ba89910ca248-kube-api-access-jbjzs\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.687541 4966 generic.go:334] "Generic (PLEG): container finished" podID="4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1" containerID="45421059253cac7584638191c6488de7cfdc0673f06d2337bddd34b4dd676cfd" exitCode=0 Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.687653 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" event={"ID":"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1","Type":"ContainerDied","Data":"45421059253cac7584638191c6488de7cfdc0673f06d2337bddd34b4dd676cfd"} Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.687694 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" event={"ID":"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1","Type":"ContainerStarted","Data":"42c9e9889c083fb26cba43e9a088d4d46cb6bf87ff546f5437e8fb25e1442fda"} Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.689799 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" event={"ID":"8f32b388-d78c-42b8-9ac2-ba89910ca248","Type":"ContainerDied","Data":"2cca30c0ee3090878148ec0d77da3e0b58672f5273a91a5979b1a88b2187a399"} Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.689839 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cca30c0ee3090878148ec0d77da3e0b58672f5273a91a5979b1a88b2187a399" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.690060 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwbrg" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.831982 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx"] Dec 17 08:45:01 crc kubenswrapper[4966]: E1217 08:45:01.832404 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f32b388-d78c-42b8-9ac2-ba89910ca248" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.832423 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f32b388-d78c-42b8-9ac2-ba89910ca248" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.832643 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f32b388-d78c-42b8-9ac2-ba89910ca248" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.833301 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.836327 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.836557 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.836712 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.836910 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.876195 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx"] Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.982859 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s25lx\" (UniqueName: \"kubernetes.io/projected/c902f130-a830-44b0-88b4-78c1e8b39081-kube-api-access-s25lx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kgzwx\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.982931 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kgzwx\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:01 crc kubenswrapper[4966]: I1217 08:45:01.984537 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kgzwx\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:02 crc kubenswrapper[4966]: I1217 08:45:02.086702 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kgzwx\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:02 crc kubenswrapper[4966]: I1217 08:45:02.087114 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kgzwx\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:02 crc kubenswrapper[4966]: I1217 08:45:02.087268 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s25lx\" (UniqueName: \"kubernetes.io/projected/c902f130-a830-44b0-88b4-78c1e8b39081-kube-api-access-s25lx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kgzwx\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:02 crc kubenswrapper[4966]: I1217 08:45:02.092809 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kgzwx\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:02 crc kubenswrapper[4966]: I1217 08:45:02.112169 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kgzwx\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:02 crc kubenswrapper[4966]: I1217 08:45:02.114919 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s25lx\" (UniqueName: \"kubernetes.io/projected/c902f130-a830-44b0-88b4-78c1e8b39081-kube-api-access-s25lx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-kgzwx\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:02 crc kubenswrapper[4966]: I1217 08:45:02.158162 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.357586 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:03 crc kubenswrapper[4966]: W1217 08:45:03.376597 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc902f130_a830_44b0_88b4_78c1e8b39081.slice/crio-1ffd72cdeb41c5a6f42c38c5643090c3c7209bc68a4b4af22d30d8c3d4cdae09 WatchSource:0}: Error finding container 1ffd72cdeb41c5a6f42c38c5643090c3c7209bc68a4b4af22d30d8c3d4cdae09: Status 404 returned error can't find the container with id 1ffd72cdeb41c5a6f42c38c5643090c3c7209bc68a4b4af22d30d8c3d4cdae09 Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.376746 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx"] Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.517814 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qct7\" (UniqueName: \"kubernetes.io/projected/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-kube-api-access-2qct7\") pod \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.518129 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-config-volume\") pod \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.518181 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-secret-volume\") pod \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\" (UID: \"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1\") " Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.519245 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-config-volume" (OuterVolumeSpecName: "config-volume") pod "4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1" (UID: "4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.526090 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1" (UID: "4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.532440 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-kube-api-access-2qct7" (OuterVolumeSpecName: "kube-api-access-2qct7") pod "4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1" (UID: "4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1"). InnerVolumeSpecName "kube-api-access-2qct7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.620691 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.621375 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.621429 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qct7\" (UniqueName: \"kubernetes.io/projected/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1-kube-api-access-2qct7\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.735940 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" event={"ID":"4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1","Type":"ContainerDied","Data":"42c9e9889c083fb26cba43e9a088d4d46cb6bf87ff546f5437e8fb25e1442fda"} Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.736303 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42c9e9889c083fb26cba43e9a088d4d46cb6bf87ff546f5437e8fb25e1442fda" Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.736196 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8" Dec 17 08:45:03 crc kubenswrapper[4966]: I1217 08:45:03.753234 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" event={"ID":"c902f130-a830-44b0-88b4-78c1e8b39081","Type":"ContainerStarted","Data":"1ffd72cdeb41c5a6f42c38c5643090c3c7209bc68a4b4af22d30d8c3d4cdae09"} Dec 17 08:45:04 crc kubenswrapper[4966]: I1217 08:45:04.763665 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" event={"ID":"c902f130-a830-44b0-88b4-78c1e8b39081","Type":"ContainerStarted","Data":"1056aa79ee241ec59a9109f442634b41565fc5046cb3118d94a5a43d9009033d"} Dec 17 08:45:04 crc kubenswrapper[4966]: I1217 08:45:04.783837 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" podStartSLOduration=3.076406482 podStartE2EDuration="3.783815454s" podCreationTimestamp="2025-12-17 08:45:01 +0000 UTC" firstStartedPulling="2025-12-17 08:45:03.388194835 +0000 UTC m=+1438.933264777" lastFinishedPulling="2025-12-17 08:45:04.095603807 +0000 UTC m=+1439.640673749" observedRunningTime="2025-12-17 08:45:04.777499092 +0000 UTC m=+1440.322569044" watchObservedRunningTime="2025-12-17 08:45:04.783815454 +0000 UTC m=+1440.328885406" Dec 17 08:45:07 crc kubenswrapper[4966]: I1217 08:45:07.804766 4966 generic.go:334] "Generic (PLEG): container finished" podID="c902f130-a830-44b0-88b4-78c1e8b39081" containerID="1056aa79ee241ec59a9109f442634b41565fc5046cb3118d94a5a43d9009033d" exitCode=0 Dec 17 08:45:07 crc kubenswrapper[4966]: I1217 08:45:07.804849 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" event={"ID":"c902f130-a830-44b0-88b4-78c1e8b39081","Type":"ContainerDied","Data":"1056aa79ee241ec59a9109f442634b41565fc5046cb3118d94a5a43d9009033d"} Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.168762 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.333570 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s25lx\" (UniqueName: \"kubernetes.io/projected/c902f130-a830-44b0-88b4-78c1e8b39081-kube-api-access-s25lx\") pod \"c902f130-a830-44b0-88b4-78c1e8b39081\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.333700 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-inventory\") pod \"c902f130-a830-44b0-88b4-78c1e8b39081\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.333802 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-ssh-key\") pod \"c902f130-a830-44b0-88b4-78c1e8b39081\" (UID: \"c902f130-a830-44b0-88b4-78c1e8b39081\") " Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.360201 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c902f130-a830-44b0-88b4-78c1e8b39081-kube-api-access-s25lx" (OuterVolumeSpecName: "kube-api-access-s25lx") pod "c902f130-a830-44b0-88b4-78c1e8b39081" (UID: "c902f130-a830-44b0-88b4-78c1e8b39081"). InnerVolumeSpecName "kube-api-access-s25lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.377188 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c902f130-a830-44b0-88b4-78c1e8b39081" (UID: "c902f130-a830-44b0-88b4-78c1e8b39081"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.401754 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-inventory" (OuterVolumeSpecName: "inventory") pod "c902f130-a830-44b0-88b4-78c1e8b39081" (UID: "c902f130-a830-44b0-88b4-78c1e8b39081"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.436824 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s25lx\" (UniqueName: \"kubernetes.io/projected/c902f130-a830-44b0-88b4-78c1e8b39081-kube-api-access-s25lx\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.436987 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.437055 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c902f130-a830-44b0-88b4-78c1e8b39081-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.823292 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" event={"ID":"c902f130-a830-44b0-88b4-78c1e8b39081","Type":"ContainerDied","Data":"1ffd72cdeb41c5a6f42c38c5643090c3c7209bc68a4b4af22d30d8c3d4cdae09"} Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.823597 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ffd72cdeb41c5a6f42c38c5643090c3c7209bc68a4b4af22d30d8c3d4cdae09" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.823345 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-kgzwx" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.925885 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh"] Dec 17 08:45:09 crc kubenswrapper[4966]: E1217 08:45:09.926274 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1" containerName="collect-profiles" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.926290 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1" containerName="collect-profiles" Dec 17 08:45:09 crc kubenswrapper[4966]: E1217 08:45:09.926322 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c902f130-a830-44b0-88b4-78c1e8b39081" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.926329 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c902f130-a830-44b0-88b4-78c1e8b39081" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.926516 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c902f130-a830-44b0-88b4-78c1e8b39081" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.926532 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1" containerName="collect-profiles" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.928790 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.932389 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.932624 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.932810 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.932971 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:45:09 crc kubenswrapper[4966]: I1217 08:45:09.943073 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh"] Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.052085 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.052191 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.052239 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qckqp\" (UniqueName: \"kubernetes.io/projected/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-kube-api-access-qckqp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.052617 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.154515 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.154589 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.154627 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qckqp\" (UniqueName: \"kubernetes.io/projected/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-kube-api-access-qckqp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.154728 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.167454 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.167464 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.169444 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.171773 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qckqp\" (UniqueName: \"kubernetes.io/projected/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-kube-api-access-qckqp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.256365 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.818626 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh"] Dec 17 08:45:10 crc kubenswrapper[4966]: I1217 08:45:10.841640 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" event={"ID":"6ef117f7-5cfa-4413-8c0c-1bb8a260749a","Type":"ContainerStarted","Data":"b2a32d37bf7fbbb41f73f04a4a14e3e83d7edbd5cf5f620bb67aa141ea285f07"} Dec 17 08:45:12 crc kubenswrapper[4966]: I1217 08:45:12.855321 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" event={"ID":"6ef117f7-5cfa-4413-8c0c-1bb8a260749a","Type":"ContainerStarted","Data":"35022e6c63c712bf1bf7fce0571e118d89efd84c84b672544aef4146d3022429"} Dec 17 08:45:16 crc kubenswrapper[4966]: I1217 08:45:16.807441 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:45:16 crc kubenswrapper[4966]: I1217 08:45:16.807950 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:45:46 crc kubenswrapper[4966]: I1217 08:45:46.807535 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:45:46 crc kubenswrapper[4966]: I1217 08:45:46.808106 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:45:46 crc kubenswrapper[4966]: I1217 08:45:46.808150 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:45:46 crc kubenswrapper[4966]: I1217 08:45:46.808816 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"682169a45fcdda5953065f4e8d118ba911e57a4b24a9e7bfd550631091e0d37e"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 08:45:46 crc kubenswrapper[4966]: I1217 08:45:46.808893 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://682169a45fcdda5953065f4e8d118ba911e57a4b24a9e7bfd550631091e0d37e" gracePeriod=600 Dec 17 08:45:47 crc kubenswrapper[4966]: I1217 08:45:47.194823 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="682169a45fcdda5953065f4e8d118ba911e57a4b24a9e7bfd550631091e0d37e" exitCode=0 Dec 17 08:45:47 crc kubenswrapper[4966]: I1217 08:45:47.194910 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"682169a45fcdda5953065f4e8d118ba911e57a4b24a9e7bfd550631091e0d37e"} Dec 17 08:45:47 crc kubenswrapper[4966]: I1217 08:45:47.195236 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b"} Dec 17 08:45:47 crc kubenswrapper[4966]: I1217 08:45:47.195268 4966 scope.go:117] "RemoveContainer" containerID="ede17bc7a2bbfa2ac454efc1ee7643bc3db24c460d6b848c803354fff3d0b2f2" Dec 17 08:45:47 crc kubenswrapper[4966]: I1217 08:45:47.251279 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" podStartSLOduration=37.062218884 podStartE2EDuration="38.251263736s" podCreationTimestamp="2025-12-17 08:45:09 +0000 UTC" firstStartedPulling="2025-12-17 08:45:10.812265967 +0000 UTC m=+1446.357335909" lastFinishedPulling="2025-12-17 08:45:12.001310819 +0000 UTC m=+1447.546380761" observedRunningTime="2025-12-17 08:45:12.884113057 +0000 UTC m=+1448.429183019" watchObservedRunningTime="2025-12-17 08:45:47.251263736 +0000 UTC m=+1482.796333668" Dec 17 08:45:48 crc kubenswrapper[4966]: I1217 08:45:48.964224 4966 scope.go:117] "RemoveContainer" containerID="d235f2e117e97cc8381b5fedfe549c5806206e82a1a991bb6b0239c9826c9d51" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.430405 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qwwx8"] Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.436599 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.451220 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwwx8"] Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.452113 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb4zx\" (UniqueName: \"kubernetes.io/projected/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-kube-api-access-zb4zx\") pod \"redhat-operators-qwwx8\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.452178 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-catalog-content\") pod \"redhat-operators-qwwx8\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.452275 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-utilities\") pod \"redhat-operators-qwwx8\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.554381 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-utilities\") pod \"redhat-operators-qwwx8\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.554519 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb4zx\" (UniqueName: \"kubernetes.io/projected/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-kube-api-access-zb4zx\") pod \"redhat-operators-qwwx8\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.554564 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-catalog-content\") pod \"redhat-operators-qwwx8\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.554944 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-utilities\") pod \"redhat-operators-qwwx8\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.555016 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-catalog-content\") pod \"redhat-operators-qwwx8\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.572472 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb4zx\" (UniqueName: \"kubernetes.io/projected/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-kube-api-access-zb4zx\") pod \"redhat-operators-qwwx8\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:14 crc kubenswrapper[4966]: I1217 08:46:14.764464 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:15 crc kubenswrapper[4966]: I1217 08:46:15.217393 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwwx8"] Dec 17 08:46:15 crc kubenswrapper[4966]: I1217 08:46:15.514692 4966 generic.go:334] "Generic (PLEG): container finished" podID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerID="52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767" exitCode=0 Dec 17 08:46:15 crc kubenswrapper[4966]: I1217 08:46:15.514738 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwwx8" event={"ID":"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc","Type":"ContainerDied","Data":"52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767"} Dec 17 08:46:15 crc kubenswrapper[4966]: I1217 08:46:15.514766 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwwx8" event={"ID":"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc","Type":"ContainerStarted","Data":"b6fce5da7570e3a7c234eb7eb89c86505dd27272e7054b65cedaa9c219cccf19"} Dec 17 08:46:16 crc kubenswrapper[4966]: I1217 08:46:16.996125 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fqcwx"] Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:16.998668 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.030128 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqcwx"] Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.034078 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nccg4\" (UniqueName: \"kubernetes.io/projected/0738366f-06ef-4e79-85bf-776ac43ac19b-kube-api-access-nccg4\") pod \"redhat-marketplace-fqcwx\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.034243 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-utilities\") pod \"redhat-marketplace-fqcwx\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.034363 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-catalog-content\") pod \"redhat-marketplace-fqcwx\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.135887 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nccg4\" (UniqueName: \"kubernetes.io/projected/0738366f-06ef-4e79-85bf-776ac43ac19b-kube-api-access-nccg4\") pod \"redhat-marketplace-fqcwx\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.135948 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-utilities\") pod \"redhat-marketplace-fqcwx\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.136023 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-catalog-content\") pod \"redhat-marketplace-fqcwx\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.136420 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-catalog-content\") pod \"redhat-marketplace-fqcwx\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.136780 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-utilities\") pod \"redhat-marketplace-fqcwx\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.156842 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nccg4\" (UniqueName: \"kubernetes.io/projected/0738366f-06ef-4e79-85bf-776ac43ac19b-kube-api-access-nccg4\") pod \"redhat-marketplace-fqcwx\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.325996 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:17 crc kubenswrapper[4966]: I1217 08:46:17.538329 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwwx8" event={"ID":"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc","Type":"ContainerStarted","Data":"98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196"} Dec 17 08:46:18 crc kubenswrapper[4966]: I1217 08:46:18.144558 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqcwx"] Dec 17 08:46:18 crc kubenswrapper[4966]: W1217 08:46:18.154543 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0738366f_06ef_4e79_85bf_776ac43ac19b.slice/crio-ac2f11bcc52b421a132b96d9548c5b423a524e671a02e10ce0dc9eb726200447 WatchSource:0}: Error finding container ac2f11bcc52b421a132b96d9548c5b423a524e671a02e10ce0dc9eb726200447: Status 404 returned error can't find the container with id ac2f11bcc52b421a132b96d9548c5b423a524e671a02e10ce0dc9eb726200447 Dec 17 08:46:18 crc kubenswrapper[4966]: I1217 08:46:18.550627 4966 generic.go:334] "Generic (PLEG): container finished" podID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerID="02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c" exitCode=0 Dec 17 08:46:18 crc kubenswrapper[4966]: I1217 08:46:18.550698 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqcwx" event={"ID":"0738366f-06ef-4e79-85bf-776ac43ac19b","Type":"ContainerDied","Data":"02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c"} Dec 17 08:46:18 crc kubenswrapper[4966]: I1217 08:46:18.551558 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqcwx" event={"ID":"0738366f-06ef-4e79-85bf-776ac43ac19b","Type":"ContainerStarted","Data":"ac2f11bcc52b421a132b96d9548c5b423a524e671a02e10ce0dc9eb726200447"} Dec 17 08:46:21 crc kubenswrapper[4966]: I1217 08:46:21.589694 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqcwx" event={"ID":"0738366f-06ef-4e79-85bf-776ac43ac19b","Type":"ContainerStarted","Data":"4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544"} Dec 17 08:46:21 crc kubenswrapper[4966]: I1217 08:46:21.593625 4966 generic.go:334] "Generic (PLEG): container finished" podID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerID="98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196" exitCode=0 Dec 17 08:46:21 crc kubenswrapper[4966]: I1217 08:46:21.593675 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwwx8" event={"ID":"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc","Type":"ContainerDied","Data":"98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196"} Dec 17 08:46:22 crc kubenswrapper[4966]: I1217 08:46:22.609619 4966 generic.go:334] "Generic (PLEG): container finished" podID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerID="4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544" exitCode=0 Dec 17 08:46:22 crc kubenswrapper[4966]: I1217 08:46:22.610314 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqcwx" event={"ID":"0738366f-06ef-4e79-85bf-776ac43ac19b","Type":"ContainerDied","Data":"4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544"} Dec 17 08:46:23 crc kubenswrapper[4966]: I1217 08:46:23.620259 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqcwx" event={"ID":"0738366f-06ef-4e79-85bf-776ac43ac19b","Type":"ContainerStarted","Data":"6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81"} Dec 17 08:46:23 crc kubenswrapper[4966]: I1217 08:46:23.623058 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwwx8" event={"ID":"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc","Type":"ContainerStarted","Data":"634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d"} Dec 17 08:46:23 crc kubenswrapper[4966]: I1217 08:46:23.649589 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fqcwx" podStartSLOduration=3.177749247 podStartE2EDuration="7.649570788s" podCreationTimestamp="2025-12-17 08:46:16 +0000 UTC" firstStartedPulling="2025-12-17 08:46:18.552727918 +0000 UTC m=+1514.097797860" lastFinishedPulling="2025-12-17 08:46:23.024549459 +0000 UTC m=+1518.569619401" observedRunningTime="2025-12-17 08:46:23.641066566 +0000 UTC m=+1519.186136508" watchObservedRunningTime="2025-12-17 08:46:23.649570788 +0000 UTC m=+1519.194640730" Dec 17 08:46:23 crc kubenswrapper[4966]: I1217 08:46:23.676727 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qwwx8" podStartSLOduration=2.6792756239999997 podStartE2EDuration="9.676707391s" podCreationTimestamp="2025-12-17 08:46:14 +0000 UTC" firstStartedPulling="2025-12-17 08:46:15.516777198 +0000 UTC m=+1511.061847140" lastFinishedPulling="2025-12-17 08:46:22.514208965 +0000 UTC m=+1518.059278907" observedRunningTime="2025-12-17 08:46:23.668477216 +0000 UTC m=+1519.213547148" watchObservedRunningTime="2025-12-17 08:46:23.676707391 +0000 UTC m=+1519.221777333" Dec 17 08:46:24 crc kubenswrapper[4966]: I1217 08:46:24.765433 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:24 crc kubenswrapper[4966]: I1217 08:46:24.766039 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:25 crc kubenswrapper[4966]: I1217 08:46:25.816850 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qwwx8" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="registry-server" probeResult="failure" output=< Dec 17 08:46:25 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:46:25 crc kubenswrapper[4966]: > Dec 17 08:46:27 crc kubenswrapper[4966]: I1217 08:46:27.326618 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:27 crc kubenswrapper[4966]: I1217 08:46:27.326917 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:27 crc kubenswrapper[4966]: I1217 08:46:27.379823 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:35 crc kubenswrapper[4966]: I1217 08:46:35.813099 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qwwx8" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="registry-server" probeResult="failure" output=< Dec 17 08:46:35 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:46:35 crc kubenswrapper[4966]: > Dec 17 08:46:37 crc kubenswrapper[4966]: I1217 08:46:37.372011 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:37 crc kubenswrapper[4966]: I1217 08:46:37.437022 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqcwx"] Dec 17 08:46:37 crc kubenswrapper[4966]: I1217 08:46:37.766936 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fqcwx" podUID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerName="registry-server" containerID="cri-o://6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81" gracePeriod=2 Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.311506 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.500279 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nccg4\" (UniqueName: \"kubernetes.io/projected/0738366f-06ef-4e79-85bf-776ac43ac19b-kube-api-access-nccg4\") pod \"0738366f-06ef-4e79-85bf-776ac43ac19b\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.500407 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-catalog-content\") pod \"0738366f-06ef-4e79-85bf-776ac43ac19b\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.500447 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-utilities\") pod \"0738366f-06ef-4e79-85bf-776ac43ac19b\" (UID: \"0738366f-06ef-4e79-85bf-776ac43ac19b\") " Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.501597 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-utilities" (OuterVolumeSpecName: "utilities") pod "0738366f-06ef-4e79-85bf-776ac43ac19b" (UID: "0738366f-06ef-4e79-85bf-776ac43ac19b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.509171 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0738366f-06ef-4e79-85bf-776ac43ac19b-kube-api-access-nccg4" (OuterVolumeSpecName: "kube-api-access-nccg4") pod "0738366f-06ef-4e79-85bf-776ac43ac19b" (UID: "0738366f-06ef-4e79-85bf-776ac43ac19b"). InnerVolumeSpecName "kube-api-access-nccg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.518722 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0738366f-06ef-4e79-85bf-776ac43ac19b" (UID: "0738366f-06ef-4e79-85bf-776ac43ac19b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.602986 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nccg4\" (UniqueName: \"kubernetes.io/projected/0738366f-06ef-4e79-85bf-776ac43ac19b-kube-api-access-nccg4\") on node \"crc\" DevicePath \"\"" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.603290 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.603357 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0738366f-06ef-4e79-85bf-776ac43ac19b-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.778968 4966 generic.go:334] "Generic (PLEG): container finished" podID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerID="6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81" exitCode=0 Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.779009 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqcwx" event={"ID":"0738366f-06ef-4e79-85bf-776ac43ac19b","Type":"ContainerDied","Data":"6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81"} Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.779039 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqcwx" event={"ID":"0738366f-06ef-4e79-85bf-776ac43ac19b","Type":"ContainerDied","Data":"ac2f11bcc52b421a132b96d9548c5b423a524e671a02e10ce0dc9eb726200447"} Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.779056 4966 scope.go:117] "RemoveContainer" containerID="6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.779067 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqcwx" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.822510 4966 scope.go:117] "RemoveContainer" containerID="4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.823248 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqcwx"] Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.846149 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqcwx"] Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.855381 4966 scope.go:117] "RemoveContainer" containerID="02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.891243 4966 scope.go:117] "RemoveContainer" containerID="6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81" Dec 17 08:46:38 crc kubenswrapper[4966]: E1217 08:46:38.892130 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81\": container with ID starting with 6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81 not found: ID does not exist" containerID="6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.892222 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81"} err="failed to get container status \"6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81\": rpc error: code = NotFound desc = could not find container \"6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81\": container with ID starting with 6b0b49a575f46be966d1567155b393074f41e6a2e11828531f25476818fb2b81 not found: ID does not exist" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.892332 4966 scope.go:117] "RemoveContainer" containerID="4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544" Dec 17 08:46:38 crc kubenswrapper[4966]: E1217 08:46:38.892677 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544\": container with ID starting with 4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544 not found: ID does not exist" containerID="4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.892705 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544"} err="failed to get container status \"4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544\": rpc error: code = NotFound desc = could not find container \"4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544\": container with ID starting with 4e0dbcf0e387313c39e819d729fe4e670c84e76ec6744be62a3f87c527411544 not found: ID does not exist" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.892723 4966 scope.go:117] "RemoveContainer" containerID="02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c" Dec 17 08:46:38 crc kubenswrapper[4966]: E1217 08:46:38.892973 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c\": container with ID starting with 02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c not found: ID does not exist" containerID="02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c" Dec 17 08:46:38 crc kubenswrapper[4966]: I1217 08:46:38.893002 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c"} err="failed to get container status \"02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c\": rpc error: code = NotFound desc = could not find container \"02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c\": container with ID starting with 02619b9c9cb6e54b23054fb9616d93a731304ad8880dafbadbb7dcd1a0a5e28c not found: ID does not exist" Dec 17 08:46:40 crc kubenswrapper[4966]: I1217 08:46:40.845403 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0738366f-06ef-4e79-85bf-776ac43ac19b" path="/var/lib/kubelet/pods/0738366f-06ef-4e79-85bf-776ac43ac19b/volumes" Dec 17 08:46:44 crc kubenswrapper[4966]: I1217 08:46:44.809977 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:44 crc kubenswrapper[4966]: I1217 08:46:44.865834 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:45 crc kubenswrapper[4966]: I1217 08:46:45.636191 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwwx8"] Dec 17 08:46:45 crc kubenswrapper[4966]: I1217 08:46:45.850050 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qwwx8" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="registry-server" containerID="cri-o://634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d" gracePeriod=2 Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.331641 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.406514 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-utilities\") pod \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.406794 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-catalog-content\") pod \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.406839 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb4zx\" (UniqueName: \"kubernetes.io/projected/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-kube-api-access-zb4zx\") pod \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\" (UID: \"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc\") " Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.407504 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-utilities" (OuterVolumeSpecName: "utilities") pod "7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" (UID: "7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.412889 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-kube-api-access-zb4zx" (OuterVolumeSpecName: "kube-api-access-zb4zx") pod "7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" (UID: "7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc"). InnerVolumeSpecName "kube-api-access-zb4zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.516132 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb4zx\" (UniqueName: \"kubernetes.io/projected/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-kube-api-access-zb4zx\") on node \"crc\" DevicePath \"\"" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.516383 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.516298 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" (UID: "7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.618275 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.864165 4966 generic.go:334] "Generic (PLEG): container finished" podID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerID="634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d" exitCode=0 Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.864221 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwwx8" event={"ID":"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc","Type":"ContainerDied","Data":"634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d"} Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.864256 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwwx8" event={"ID":"7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc","Type":"ContainerDied","Data":"b6fce5da7570e3a7c234eb7eb89c86505dd27272e7054b65cedaa9c219cccf19"} Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.864280 4966 scope.go:117] "RemoveContainer" containerID="634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.864477 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwwx8" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.901287 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwwx8"] Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.921246 4966 scope.go:117] "RemoveContainer" containerID="98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196" Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.924352 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qwwx8"] Dec 17 08:46:46 crc kubenswrapper[4966]: I1217 08:46:46.950968 4966 scope.go:117] "RemoveContainer" containerID="52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767" Dec 17 08:46:47 crc kubenswrapper[4966]: I1217 08:46:47.002488 4966 scope.go:117] "RemoveContainer" containerID="634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d" Dec 17 08:46:47 crc kubenswrapper[4966]: E1217 08:46:47.011073 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d\": container with ID starting with 634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d not found: ID does not exist" containerID="634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d" Dec 17 08:46:47 crc kubenswrapper[4966]: I1217 08:46:47.011124 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d"} err="failed to get container status \"634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d\": rpc error: code = NotFound desc = could not find container \"634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d\": container with ID starting with 634e4561a82c2dc594af144a89d836fbebb2a896ef5c39c45e53a8af5a99f33d not found: ID does not exist" Dec 17 08:46:47 crc kubenswrapper[4966]: I1217 08:46:47.011157 4966 scope.go:117] "RemoveContainer" containerID="98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196" Dec 17 08:46:47 crc kubenswrapper[4966]: E1217 08:46:47.014799 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196\": container with ID starting with 98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196 not found: ID does not exist" containerID="98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196" Dec 17 08:46:47 crc kubenswrapper[4966]: I1217 08:46:47.014836 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196"} err="failed to get container status \"98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196\": rpc error: code = NotFound desc = could not find container \"98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196\": container with ID starting with 98bab6d032156002eda0598d1346eb247ee03edb5b6b0efe628d306ec562e196 not found: ID does not exist" Dec 17 08:46:47 crc kubenswrapper[4966]: I1217 08:46:47.014864 4966 scope.go:117] "RemoveContainer" containerID="52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767" Dec 17 08:46:47 crc kubenswrapper[4966]: E1217 08:46:47.015541 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767\": container with ID starting with 52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767 not found: ID does not exist" containerID="52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767" Dec 17 08:46:47 crc kubenswrapper[4966]: I1217 08:46:47.015571 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767"} err="failed to get container status \"52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767\": rpc error: code = NotFound desc = could not find container \"52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767\": container with ID starting with 52370089477bcc06edb429bd7105bf3597aa8798506e5ee63cbc2c51ddba1767 not found: ID does not exist" Dec 17 08:46:48 crc kubenswrapper[4966]: I1217 08:46:48.840146 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" path="/var/lib/kubelet/pods/7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc/volumes" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.590643 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pzq6n"] Dec 17 08:46:56 crc kubenswrapper[4966]: E1217 08:46:56.591721 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="registry-server" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.591737 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="registry-server" Dec 17 08:46:56 crc kubenswrapper[4966]: E1217 08:46:56.591760 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="extract-utilities" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.591768 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="extract-utilities" Dec 17 08:46:56 crc kubenswrapper[4966]: E1217 08:46:56.591776 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="extract-content" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.591782 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="extract-content" Dec 17 08:46:56 crc kubenswrapper[4966]: E1217 08:46:56.591809 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerName="extract-content" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.591815 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerName="extract-content" Dec 17 08:46:56 crc kubenswrapper[4966]: E1217 08:46:56.591823 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerName="registry-server" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.591828 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerName="registry-server" Dec 17 08:46:56 crc kubenswrapper[4966]: E1217 08:46:56.591844 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerName="extract-utilities" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.591850 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerName="extract-utilities" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.592079 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0738366f-06ef-4e79-85bf-776ac43ac19b" containerName="registry-server" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.592094 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc1c4c1-132c-4788-8f8f-44ee7b67fbbc" containerName="registry-server" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.593525 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.600281 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pzq6n"] Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.747101 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-utilities\") pod \"certified-operators-pzq6n\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.747472 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nvlm\" (UniqueName: \"kubernetes.io/projected/aa8739b4-e768-49b4-ad32-682731b8d100-kube-api-access-4nvlm\") pod \"certified-operators-pzq6n\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.747556 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-catalog-content\") pod \"certified-operators-pzq6n\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.849897 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-catalog-content\") pod \"certified-operators-pzq6n\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.850012 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-utilities\") pod \"certified-operators-pzq6n\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.850094 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nvlm\" (UniqueName: \"kubernetes.io/projected/aa8739b4-e768-49b4-ad32-682731b8d100-kube-api-access-4nvlm\") pod \"certified-operators-pzq6n\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.851559 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-catalog-content\") pod \"certified-operators-pzq6n\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.851929 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-utilities\") pod \"certified-operators-pzq6n\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.878195 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nvlm\" (UniqueName: \"kubernetes.io/projected/aa8739b4-e768-49b4-ad32-682731b8d100-kube-api-access-4nvlm\") pod \"certified-operators-pzq6n\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:56 crc kubenswrapper[4966]: I1217 08:46:56.927136 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:46:57 crc kubenswrapper[4966]: I1217 08:46:57.251243 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pzq6n"] Dec 17 08:46:57 crc kubenswrapper[4966]: I1217 08:46:57.976397 4966 generic.go:334] "Generic (PLEG): container finished" podID="aa8739b4-e768-49b4-ad32-682731b8d100" containerID="18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d" exitCode=0 Dec 17 08:46:57 crc kubenswrapper[4966]: I1217 08:46:57.976638 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzq6n" event={"ID":"aa8739b4-e768-49b4-ad32-682731b8d100","Type":"ContainerDied","Data":"18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d"} Dec 17 08:46:57 crc kubenswrapper[4966]: I1217 08:46:57.976797 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzq6n" event={"ID":"aa8739b4-e768-49b4-ad32-682731b8d100","Type":"ContainerStarted","Data":"02c2c16b3def7fac9b59c1bf6de6a11773d367e56c7b755ef98926abbd0390f4"} Dec 17 08:46:57 crc kubenswrapper[4966]: I1217 08:46:57.980814 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 08:47:00 crc kubenswrapper[4966]: I1217 08:47:00.027274 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzq6n" event={"ID":"aa8739b4-e768-49b4-ad32-682731b8d100","Type":"ContainerStarted","Data":"8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2"} Dec 17 08:47:01 crc kubenswrapper[4966]: I1217 08:47:01.037715 4966 generic.go:334] "Generic (PLEG): container finished" podID="aa8739b4-e768-49b4-ad32-682731b8d100" containerID="8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2" exitCode=0 Dec 17 08:47:01 crc kubenswrapper[4966]: I1217 08:47:01.037819 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzq6n" event={"ID":"aa8739b4-e768-49b4-ad32-682731b8d100","Type":"ContainerDied","Data":"8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2"} Dec 17 08:47:02 crc kubenswrapper[4966]: I1217 08:47:02.049412 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzq6n" event={"ID":"aa8739b4-e768-49b4-ad32-682731b8d100","Type":"ContainerStarted","Data":"d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff"} Dec 17 08:47:02 crc kubenswrapper[4966]: I1217 08:47:02.071119 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pzq6n" podStartSLOduration=2.452002368 podStartE2EDuration="6.071098472s" podCreationTimestamp="2025-12-17 08:46:56 +0000 UTC" firstStartedPulling="2025-12-17 08:46:57.980213388 +0000 UTC m=+1553.525283340" lastFinishedPulling="2025-12-17 08:47:01.599309502 +0000 UTC m=+1557.144379444" observedRunningTime="2025-12-17 08:47:02.066493987 +0000 UTC m=+1557.611563939" watchObservedRunningTime="2025-12-17 08:47:02.071098472 +0000 UTC m=+1557.616168434" Dec 17 08:47:06 crc kubenswrapper[4966]: I1217 08:47:06.927537 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:47:06 crc kubenswrapper[4966]: I1217 08:47:06.929158 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:47:06 crc kubenswrapper[4966]: I1217 08:47:06.981660 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:47:07 crc kubenswrapper[4966]: I1217 08:47:07.136525 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:47:07 crc kubenswrapper[4966]: I1217 08:47:07.237335 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pzq6n"] Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.107792 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pzq6n" podUID="aa8739b4-e768-49b4-ad32-682731b8d100" containerName="registry-server" containerID="cri-o://d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff" gracePeriod=2 Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.561052 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.635562 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-catalog-content\") pod \"aa8739b4-e768-49b4-ad32-682731b8d100\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.635735 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nvlm\" (UniqueName: \"kubernetes.io/projected/aa8739b4-e768-49b4-ad32-682731b8d100-kube-api-access-4nvlm\") pod \"aa8739b4-e768-49b4-ad32-682731b8d100\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.635799 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-utilities\") pod \"aa8739b4-e768-49b4-ad32-682731b8d100\" (UID: \"aa8739b4-e768-49b4-ad32-682731b8d100\") " Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.636715 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-utilities" (OuterVolumeSpecName: "utilities") pod "aa8739b4-e768-49b4-ad32-682731b8d100" (UID: "aa8739b4-e768-49b4-ad32-682731b8d100"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.651235 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa8739b4-e768-49b4-ad32-682731b8d100-kube-api-access-4nvlm" (OuterVolumeSpecName: "kube-api-access-4nvlm") pod "aa8739b4-e768-49b4-ad32-682731b8d100" (UID: "aa8739b4-e768-49b4-ad32-682731b8d100"). InnerVolumeSpecName "kube-api-access-4nvlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.688498 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa8739b4-e768-49b4-ad32-682731b8d100" (UID: "aa8739b4-e768-49b4-ad32-682731b8d100"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.737855 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.737919 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nvlm\" (UniqueName: \"kubernetes.io/projected/aa8739b4-e768-49b4-ad32-682731b8d100-kube-api-access-4nvlm\") on node \"crc\" DevicePath \"\"" Dec 17 08:47:09 crc kubenswrapper[4966]: I1217 08:47:09.737935 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa8739b4-e768-49b4-ad32-682731b8d100-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.118955 4966 generic.go:334] "Generic (PLEG): container finished" podID="aa8739b4-e768-49b4-ad32-682731b8d100" containerID="d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff" exitCode=0 Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.119003 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzq6n" event={"ID":"aa8739b4-e768-49b4-ad32-682731b8d100","Type":"ContainerDied","Data":"d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff"} Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.119021 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzq6n" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.119034 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzq6n" event={"ID":"aa8739b4-e768-49b4-ad32-682731b8d100","Type":"ContainerDied","Data":"02c2c16b3def7fac9b59c1bf6de6a11773d367e56c7b755ef98926abbd0390f4"} Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.119054 4966 scope.go:117] "RemoveContainer" containerID="d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.163415 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pzq6n"] Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.168755 4966 scope.go:117] "RemoveContainer" containerID="8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.174003 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pzq6n"] Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.192053 4966 scope.go:117] "RemoveContainer" containerID="18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.237580 4966 scope.go:117] "RemoveContainer" containerID="d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff" Dec 17 08:47:10 crc kubenswrapper[4966]: E1217 08:47:10.238211 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff\": container with ID starting with d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff not found: ID does not exist" containerID="d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.238266 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff"} err="failed to get container status \"d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff\": rpc error: code = NotFound desc = could not find container \"d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff\": container with ID starting with d654139d75bdc741deb5cec346b2139b657160e741622bb19d270de0aed007ff not found: ID does not exist" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.238299 4966 scope.go:117] "RemoveContainer" containerID="8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2" Dec 17 08:47:10 crc kubenswrapper[4966]: E1217 08:47:10.238665 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2\": container with ID starting with 8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2 not found: ID does not exist" containerID="8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.238705 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2"} err="failed to get container status \"8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2\": rpc error: code = NotFound desc = could not find container \"8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2\": container with ID starting with 8619b50722f3bed5389dacf2200ad315a9c600f89710a6f0be044e051c8b7bd2 not found: ID does not exist" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.238734 4966 scope.go:117] "RemoveContainer" containerID="18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d" Dec 17 08:47:10 crc kubenswrapper[4966]: E1217 08:47:10.239088 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d\": container with ID starting with 18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d not found: ID does not exist" containerID="18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.239120 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d"} err="failed to get container status \"18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d\": rpc error: code = NotFound desc = could not find container \"18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d\": container with ID starting with 18facb279cbd584089dc92a6dc358a0bdf6f30bdcd05f31128246192db6c106d not found: ID does not exist" Dec 17 08:47:10 crc kubenswrapper[4966]: I1217 08:47:10.841465 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa8739b4-e768-49b4-ad32-682731b8d100" path="/var/lib/kubelet/pods/aa8739b4-e768-49b4-ad32-682731b8d100/volumes" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.340421 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cgdzl"] Dec 17 08:47:56 crc kubenswrapper[4966]: E1217 08:47:56.341284 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa8739b4-e768-49b4-ad32-682731b8d100" containerName="registry-server" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.341301 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa8739b4-e768-49b4-ad32-682731b8d100" containerName="registry-server" Dec 17 08:47:56 crc kubenswrapper[4966]: E1217 08:47:56.341321 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa8739b4-e768-49b4-ad32-682731b8d100" containerName="extract-utilities" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.341327 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa8739b4-e768-49b4-ad32-682731b8d100" containerName="extract-utilities" Dec 17 08:47:56 crc kubenswrapper[4966]: E1217 08:47:56.341354 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa8739b4-e768-49b4-ad32-682731b8d100" containerName="extract-content" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.341362 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa8739b4-e768-49b4-ad32-682731b8d100" containerName="extract-content" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.341555 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa8739b4-e768-49b4-ad32-682731b8d100" containerName="registry-server" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.345530 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.374720 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgdzl"] Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.415033 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-catalog-content\") pod \"community-operators-cgdzl\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.415144 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-utilities\") pod \"community-operators-cgdzl\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.415235 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbc6g\" (UniqueName: \"kubernetes.io/projected/c7dff045-586c-4bcd-a312-808d969ea508-kube-api-access-sbc6g\") pod \"community-operators-cgdzl\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.516787 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbc6g\" (UniqueName: \"kubernetes.io/projected/c7dff045-586c-4bcd-a312-808d969ea508-kube-api-access-sbc6g\") pod \"community-operators-cgdzl\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.516901 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-catalog-content\") pod \"community-operators-cgdzl\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.516958 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-utilities\") pod \"community-operators-cgdzl\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.517420 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-catalog-content\") pod \"community-operators-cgdzl\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.517472 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-utilities\") pod \"community-operators-cgdzl\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.537222 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbc6g\" (UniqueName: \"kubernetes.io/projected/c7dff045-586c-4bcd-a312-808d969ea508-kube-api-access-sbc6g\") pod \"community-operators-cgdzl\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:56 crc kubenswrapper[4966]: I1217 08:47:56.668090 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:47:57 crc kubenswrapper[4966]: I1217 08:47:57.136998 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgdzl"] Dec 17 08:47:57 crc kubenswrapper[4966]: I1217 08:47:57.685043 4966 generic.go:334] "Generic (PLEG): container finished" podID="c7dff045-586c-4bcd-a312-808d969ea508" containerID="e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001" exitCode=0 Dec 17 08:47:57 crc kubenswrapper[4966]: I1217 08:47:57.685211 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdzl" event={"ID":"c7dff045-586c-4bcd-a312-808d969ea508","Type":"ContainerDied","Data":"e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001"} Dec 17 08:47:57 crc kubenswrapper[4966]: I1217 08:47:57.685328 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdzl" event={"ID":"c7dff045-586c-4bcd-a312-808d969ea508","Type":"ContainerStarted","Data":"0a3d44ba27e6e9b9d7cb08d76c4d0d928fe09ac7861e08673c45eb95ef4ec7d5"} Dec 17 08:47:59 crc kubenswrapper[4966]: I1217 08:47:59.708757 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdzl" event={"ID":"c7dff045-586c-4bcd-a312-808d969ea508","Type":"ContainerStarted","Data":"d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6"} Dec 17 08:48:00 crc kubenswrapper[4966]: I1217 08:48:00.725558 4966 generic.go:334] "Generic (PLEG): container finished" podID="c7dff045-586c-4bcd-a312-808d969ea508" containerID="d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6" exitCode=0 Dec 17 08:48:00 crc kubenswrapper[4966]: I1217 08:48:00.725633 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdzl" event={"ID":"c7dff045-586c-4bcd-a312-808d969ea508","Type":"ContainerDied","Data":"d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6"} Dec 17 08:48:01 crc kubenswrapper[4966]: I1217 08:48:01.736862 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdzl" event={"ID":"c7dff045-586c-4bcd-a312-808d969ea508","Type":"ContainerStarted","Data":"04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36"} Dec 17 08:48:01 crc kubenswrapper[4966]: I1217 08:48:01.755556 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cgdzl" podStartSLOduration=1.969461527 podStartE2EDuration="5.755540118s" podCreationTimestamp="2025-12-17 08:47:56 +0000 UTC" firstStartedPulling="2025-12-17 08:47:57.687197319 +0000 UTC m=+1613.232267261" lastFinishedPulling="2025-12-17 08:48:01.47327591 +0000 UTC m=+1617.018345852" observedRunningTime="2025-12-17 08:48:01.751926119 +0000 UTC m=+1617.296996071" watchObservedRunningTime="2025-12-17 08:48:01.755540118 +0000 UTC m=+1617.300610060" Dec 17 08:48:02 crc kubenswrapper[4966]: I1217 08:48:02.071459 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-0937-account-create-update-bq6tm"] Dec 17 08:48:02 crc kubenswrapper[4966]: I1217 08:48:02.085770 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-6l529"] Dec 17 08:48:02 crc kubenswrapper[4966]: I1217 08:48:02.098370 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-6l529"] Dec 17 08:48:02 crc kubenswrapper[4966]: I1217 08:48:02.110771 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-0937-account-create-update-bq6tm"] Dec 17 08:48:02 crc kubenswrapper[4966]: I1217 08:48:02.859093 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3153dd6b-a60e-4da6-8463-a6e7b86c0277" path="/var/lib/kubelet/pods/3153dd6b-a60e-4da6-8463-a6e7b86c0277/volumes" Dec 17 08:48:02 crc kubenswrapper[4966]: I1217 08:48:02.861850 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d6c04f-29ad-41b5-a6b7-a8868cf02155" path="/var/lib/kubelet/pods/84d6c04f-29ad-41b5-a6b7-a8868cf02155/volumes" Dec 17 08:48:06 crc kubenswrapper[4966]: I1217 08:48:06.668893 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:48:06 crc kubenswrapper[4966]: I1217 08:48:06.669347 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:48:06 crc kubenswrapper[4966]: I1217 08:48:06.713003 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:48:06 crc kubenswrapper[4966]: I1217 08:48:06.827106 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:48:06 crc kubenswrapper[4966]: I1217 08:48:06.950363 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgdzl"] Dec 17 08:48:08 crc kubenswrapper[4966]: I1217 08:48:08.060518 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-c68tj"] Dec 17 08:48:08 crc kubenswrapper[4966]: I1217 08:48:08.074217 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-znk6r"] Dec 17 08:48:08 crc kubenswrapper[4966]: I1217 08:48:08.085310 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-znk6r"] Dec 17 08:48:08 crc kubenswrapper[4966]: I1217 08:48:08.097254 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-c68tj"] Dec 17 08:48:08 crc kubenswrapper[4966]: I1217 08:48:08.808787 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cgdzl" podUID="c7dff045-586c-4bcd-a312-808d969ea508" containerName="registry-server" containerID="cri-o://04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36" gracePeriod=2 Dec 17 08:48:08 crc kubenswrapper[4966]: I1217 08:48:08.849566 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d6079f-863e-48f6-a8d8-66ffda331c89" path="/var/lib/kubelet/pods/24d6079f-863e-48f6-a8d8-66ffda331c89/volumes" Dec 17 08:48:08 crc kubenswrapper[4966]: I1217 08:48:08.851125 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183" path="/var/lib/kubelet/pods/8d1eca5f-c69b-4ac6-9dfe-2ed6a9958183/volumes" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.796190 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.845377 4966 generic.go:334] "Generic (PLEG): container finished" podID="c7dff045-586c-4bcd-a312-808d969ea508" containerID="04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36" exitCode=0 Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.845662 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdzl" event={"ID":"c7dff045-586c-4bcd-a312-808d969ea508","Type":"ContainerDied","Data":"04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36"} Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.845768 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdzl" event={"ID":"c7dff045-586c-4bcd-a312-808d969ea508","Type":"ContainerDied","Data":"0a3d44ba27e6e9b9d7cb08d76c4d0d928fe09ac7861e08673c45eb95ef4ec7d5"} Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.845857 4966 scope.go:117] "RemoveContainer" containerID="04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.846342 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgdzl" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.873963 4966 scope.go:117] "RemoveContainer" containerID="d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.891028 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-utilities\") pod \"c7dff045-586c-4bcd-a312-808d969ea508\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.891299 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbc6g\" (UniqueName: \"kubernetes.io/projected/c7dff045-586c-4bcd-a312-808d969ea508-kube-api-access-sbc6g\") pod \"c7dff045-586c-4bcd-a312-808d969ea508\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.891383 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-catalog-content\") pod \"c7dff045-586c-4bcd-a312-808d969ea508\" (UID: \"c7dff045-586c-4bcd-a312-808d969ea508\") " Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.895755 4966 scope.go:117] "RemoveContainer" containerID="e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.896539 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-utilities" (OuterVolumeSpecName: "utilities") pod "c7dff045-586c-4bcd-a312-808d969ea508" (UID: "c7dff045-586c-4bcd-a312-808d969ea508"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.903308 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7dff045-586c-4bcd-a312-808d969ea508-kube-api-access-sbc6g" (OuterVolumeSpecName: "kube-api-access-sbc6g") pod "c7dff045-586c-4bcd-a312-808d969ea508" (UID: "c7dff045-586c-4bcd-a312-808d969ea508"). InnerVolumeSpecName "kube-api-access-sbc6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.954130 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7dff045-586c-4bcd-a312-808d969ea508" (UID: "c7dff045-586c-4bcd-a312-808d969ea508"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.995992 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbc6g\" (UniqueName: \"kubernetes.io/projected/c7dff045-586c-4bcd-a312-808d969ea508-kube-api-access-sbc6g\") on node \"crc\" DevicePath \"\"" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.996017 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:48:09 crc kubenswrapper[4966]: I1217 08:48:09.996026 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7dff045-586c-4bcd-a312-808d969ea508-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.002224 4966 scope.go:117] "RemoveContainer" containerID="04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36" Dec 17 08:48:10 crc kubenswrapper[4966]: E1217 08:48:10.002699 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36\": container with ID starting with 04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36 not found: ID does not exist" containerID="04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36" Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.002741 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36"} err="failed to get container status \"04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36\": rpc error: code = NotFound desc = could not find container \"04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36\": container with ID starting with 04a085632e9a3638e9c99dc2055a942635536479cd874808363a7f4cf8c71f36 not found: ID does not exist" Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.002768 4966 scope.go:117] "RemoveContainer" containerID="d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6" Dec 17 08:48:10 crc kubenswrapper[4966]: E1217 08:48:10.003151 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6\": container with ID starting with d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6 not found: ID does not exist" containerID="d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6" Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.003189 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6"} err="failed to get container status \"d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6\": rpc error: code = NotFound desc = could not find container \"d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6\": container with ID starting with d8581317ac8925c302aca4e5ab94ffed8746c95784481e1fc095563a816c8da6 not found: ID does not exist" Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.003216 4966 scope.go:117] "RemoveContainer" containerID="e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001" Dec 17 08:48:10 crc kubenswrapper[4966]: E1217 08:48:10.003666 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001\": container with ID starting with e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001 not found: ID does not exist" containerID="e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001" Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.003695 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001"} err="failed to get container status \"e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001\": rpc error: code = NotFound desc = could not find container \"e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001\": container with ID starting with e03fb331c10ae53ce34957a3c5d119a69faadb4ecd6d25952d76f657b4fd5001 not found: ID does not exist" Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.039472 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-382e-account-create-update-tvqcb"] Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.049739 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-097c-account-create-update-sg2zd"] Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.060618 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-097c-account-create-update-sg2zd"] Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.069363 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-382e-account-create-update-tvqcb"] Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.182002 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgdzl"] Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.190660 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cgdzl"] Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.857193 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00699bc6-288b-4a56-92f4-280d1371f224" path="/var/lib/kubelet/pods/00699bc6-288b-4a56-92f4-280d1371f224/volumes" Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.861958 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a603f92f-57a4-474a-b8e8-175f528a7a08" path="/var/lib/kubelet/pods/a603f92f-57a4-474a-b8e8-175f528a7a08/volumes" Dec 17 08:48:10 crc kubenswrapper[4966]: I1217 08:48:10.864204 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7dff045-586c-4bcd-a312-808d969ea508" path="/var/lib/kubelet/pods/c7dff045-586c-4bcd-a312-808d969ea508/volumes" Dec 17 08:48:16 crc kubenswrapper[4966]: I1217 08:48:16.807641 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:48:16 crc kubenswrapper[4966]: I1217 08:48:16.808273 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:48:43 crc kubenswrapper[4966]: I1217 08:48:43.046606 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-vp2ln"] Dec 17 08:48:43 crc kubenswrapper[4966]: I1217 08:48:43.057537 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-vp2ln"] Dec 17 08:48:44 crc kubenswrapper[4966]: I1217 08:48:44.842309 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06" path="/var/lib/kubelet/pods/99f1eaf1-29a7-4fa6-bd71-19e3b0e49b06/volumes" Dec 17 08:48:46 crc kubenswrapper[4966]: I1217 08:48:46.807920 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:48:46 crc kubenswrapper[4966]: I1217 08:48:46.808257 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:48:49 crc kubenswrapper[4966]: I1217 08:48:49.120141 4966 scope.go:117] "RemoveContainer" containerID="399e5d4c8d544e1fd09d3550c8c61046fd1485a2660334ccd4ae7f4d6f2e042c" Dec 17 08:48:49 crc kubenswrapper[4966]: I1217 08:48:49.164069 4966 scope.go:117] "RemoveContainer" containerID="dcc7737790f864cddd18451c40f96b905d2b29ea5db9fa864afce06c80e6fa5d" Dec 17 08:48:49 crc kubenswrapper[4966]: I1217 08:48:49.235804 4966 scope.go:117] "RemoveContainer" containerID="24696c1cc38752d3b8fb935616ad75818333cf6e8043d3d2bf8a31fc814d5382" Dec 17 08:48:49 crc kubenswrapper[4966]: I1217 08:48:49.294865 4966 scope.go:117] "RemoveContainer" containerID="dd00a55a88218f1f7dc9298b73c2400bce6f4f0fc57cc40e25643031b577826b" Dec 17 08:48:49 crc kubenswrapper[4966]: I1217 08:48:49.327499 4966 scope.go:117] "RemoveContainer" containerID="76a94e47d47be924f17c05bfa6c88dbcf96fbbfe725c870f8d3fd62d97ba32ed" Dec 17 08:48:49 crc kubenswrapper[4966]: I1217 08:48:49.369057 4966 scope.go:117] "RemoveContainer" containerID="730743ecd154551f18d4f8797663937b0d35e8a91fb5e32a07c62d2181f4c277" Dec 17 08:48:49 crc kubenswrapper[4966]: I1217 08:48:49.416232 4966 scope.go:117] "RemoveContainer" containerID="fd04fc0fe9a3d320c1a8971bf2f586b5158d04fd71966ab3ef1a5dcf81409e6b" Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.040460 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-jpx62"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.053078 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-952b-account-create-update-p5444"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.066860 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d3cd-account-create-update-4ccdt"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.076109 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-jpx62"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.085494 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-f803-account-create-update-p88mq"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.094230 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d3cd-account-create-update-4ccdt"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.104128 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-f803-account-create-update-p88mq"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.112404 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-952b-account-create-update-p5444"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.120109 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-rjxqq"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.127615 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-rjxqq"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.135224 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-e584-account-create-update-hk75s"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.143452 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-e584-account-create-update-hk75s"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.151024 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-z57pq"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.158800 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-fp4tg"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.166569 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-fp4tg"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.176342 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-z57pq"] Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.843976 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23a69382-8e08-47cf-a9aa-bc254229aa21" path="/var/lib/kubelet/pods/23a69382-8e08-47cf-a9aa-bc254229aa21/volumes" Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.845648 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f4b0470-0072-4fdc-8a49-33b8c5827ba0" path="/var/lib/kubelet/pods/4f4b0470-0072-4fdc-8a49-33b8c5827ba0/volumes" Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.847124 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="552f3591-2913-4bbb-bf5b-79ea4ba40778" path="/var/lib/kubelet/pods/552f3591-2913-4bbb-bf5b-79ea4ba40778/volumes" Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.848023 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="771eea8c-0e4e-471c-be4b-3f18e3d91adb" path="/var/lib/kubelet/pods/771eea8c-0e4e-471c-be4b-3f18e3d91adb/volumes" Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.849356 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="937dc9ee-a952-4496-9659-b4cc9299018a" path="/var/lib/kubelet/pods/937dc9ee-a952-4496-9659-b4cc9299018a/volumes" Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.850304 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a952c157-89d5-4bb4-a624-b5b43babd9e0" path="/var/lib/kubelet/pods/a952c157-89d5-4bb4-a624-b5b43babd9e0/volumes" Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.851241 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be459859-4e2c-4548-a544-39bf15a731b0" path="/var/lib/kubelet/pods/be459859-4e2c-4548-a544-39bf15a731b0/volumes" Dec 17 08:48:50 crc kubenswrapper[4966]: I1217 08:48:50.852197 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f70fc4b4-6cf8-4852-827a-d2939c1b656a" path="/var/lib/kubelet/pods/f70fc4b4-6cf8-4852-827a-d2939c1b656a/volumes" Dec 17 08:48:56 crc kubenswrapper[4966]: I1217 08:48:56.034077 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-s8bmz"] Dec 17 08:48:56 crc kubenswrapper[4966]: I1217 08:48:56.046354 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-s8bmz"] Dec 17 08:48:56 crc kubenswrapper[4966]: I1217 08:48:56.842776 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8603b503-2df3-4a05-84e7-024e63d2b455" path="/var/lib/kubelet/pods/8603b503-2df3-4a05-84e7-024e63d2b455/volumes" Dec 17 08:49:16 crc kubenswrapper[4966]: I1217 08:49:16.808045 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:49:16 crc kubenswrapper[4966]: I1217 08:49:16.808666 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:49:16 crc kubenswrapper[4966]: I1217 08:49:16.808717 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:49:16 crc kubenswrapper[4966]: I1217 08:49:16.809478 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 08:49:16 crc kubenswrapper[4966]: I1217 08:49:16.809539 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" gracePeriod=600 Dec 17 08:49:16 crc kubenswrapper[4966]: E1217 08:49:16.950973 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:49:17 crc kubenswrapper[4966]: I1217 08:49:17.569959 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" exitCode=0 Dec 17 08:49:17 crc kubenswrapper[4966]: I1217 08:49:17.570055 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b"} Dec 17 08:49:17 crc kubenswrapper[4966]: I1217 08:49:17.570395 4966 scope.go:117] "RemoveContainer" containerID="682169a45fcdda5953065f4e8d118ba911e57a4b24a9e7bfd550631091e0d37e" Dec 17 08:49:17 crc kubenswrapper[4966]: I1217 08:49:17.571250 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:49:17 crc kubenswrapper[4966]: E1217 08:49:17.571818 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:49:28 crc kubenswrapper[4966]: I1217 08:49:28.830574 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:49:28 crc kubenswrapper[4966]: E1217 08:49:28.831265 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:49:35 crc kubenswrapper[4966]: I1217 08:49:35.041927 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-fhj4g"] Dec 17 08:49:35 crc kubenswrapper[4966]: I1217 08:49:35.050034 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-fhj4g"] Dec 17 08:49:36 crc kubenswrapper[4966]: I1217 08:49:36.854957 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62" path="/var/lib/kubelet/pods/b52f1e67-1d11-4eb6-bf67-ca0f3a5c1c62/volumes" Dec 17 08:49:42 crc kubenswrapper[4966]: I1217 08:49:42.831061 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:49:42 crc kubenswrapper[4966]: E1217 08:49:42.831866 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.603415 4966 scope.go:117] "RemoveContainer" containerID="9e2a12bca64d76f2997a3d267843d0bbdfedf734b641afb2cc96aaf0816e82a7" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.629664 4966 scope.go:117] "RemoveContainer" containerID="8adae0b53e1d6b72b2493930742ba86d80089e13ff59ca29f7d345e4edc0c72f" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.686431 4966 scope.go:117] "RemoveContainer" containerID="79f04cd102f2163311766ff25fb8b27cb5227115c3882b44f13422cee82b6b26" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.742374 4966 scope.go:117] "RemoveContainer" containerID="f79229927e2f10f0d9db670083d2f547170bbb349303724473992f2a2dfec5de" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.787847 4966 scope.go:117] "RemoveContainer" containerID="1808d23a4649f555ebc6d720fd7f0d7a47c89d9322352250bdbf4bef4a49fa32" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.829305 4966 scope.go:117] "RemoveContainer" containerID="9c2e22d6898f96ac40e0ae9f555d1a0fd0ce5e52652f637dd40568d743b52828" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.881546 4966 scope.go:117] "RemoveContainer" containerID="44520bfad7d211b8f12aaaf0e2ff9e26735d211290d564e51b91d0c9641fe6fe" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.899460 4966 scope.go:117] "RemoveContainer" containerID="bd94580d0d11729b13ffc6442cf457c306f3630648f4ea98bfa3a1e4d4d84037" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.921698 4966 scope.go:117] "RemoveContainer" containerID="4eb0ce01ae7d7c0e690b982e63d34832ef0d648791fa202b9c8dfbcad882c051" Dec 17 08:49:49 crc kubenswrapper[4966]: I1217 08:49:49.947270 4966 scope.go:117] "RemoveContainer" containerID="bc37c940d29882f401364b24328941b8b0b13f877bc7f4674dd8a643e6fd34a4" Dec 17 08:49:51 crc kubenswrapper[4966]: I1217 08:49:51.941323 4966 generic.go:334] "Generic (PLEG): container finished" podID="6ef117f7-5cfa-4413-8c0c-1bb8a260749a" containerID="35022e6c63c712bf1bf7fce0571e118d89efd84c84b672544aef4146d3022429" exitCode=0 Dec 17 08:49:51 crc kubenswrapper[4966]: I1217 08:49:51.941414 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" event={"ID":"6ef117f7-5cfa-4413-8c0c-1bb8a260749a","Type":"ContainerDied","Data":"35022e6c63c712bf1bf7fce0571e118d89efd84c84b672544aef4146d3022429"} Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.605404 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.766901 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-bootstrap-combined-ca-bundle\") pod \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.767096 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-inventory\") pod \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.767150 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qckqp\" (UniqueName: \"kubernetes.io/projected/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-kube-api-access-qckqp\") pod \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.767245 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-ssh-key\") pod \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\" (UID: \"6ef117f7-5cfa-4413-8c0c-1bb8a260749a\") " Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.773098 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-kube-api-access-qckqp" (OuterVolumeSpecName: "kube-api-access-qckqp") pod "6ef117f7-5cfa-4413-8c0c-1bb8a260749a" (UID: "6ef117f7-5cfa-4413-8c0c-1bb8a260749a"). InnerVolumeSpecName "kube-api-access-qckqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.789946 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6ef117f7-5cfa-4413-8c0c-1bb8a260749a" (UID: "6ef117f7-5cfa-4413-8c0c-1bb8a260749a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.816231 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6ef117f7-5cfa-4413-8c0c-1bb8a260749a" (UID: "6ef117f7-5cfa-4413-8c0c-1bb8a260749a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.816350 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-inventory" (OuterVolumeSpecName: "inventory") pod "6ef117f7-5cfa-4413-8c0c-1bb8a260749a" (UID: "6ef117f7-5cfa-4413-8c0c-1bb8a260749a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.869245 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.869460 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qckqp\" (UniqueName: \"kubernetes.io/projected/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-kube-api-access-qckqp\") on node \"crc\" DevicePath \"\"" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.869483 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.869494 4966 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ef117f7-5cfa-4413-8c0c-1bb8a260749a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.961697 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" event={"ID":"6ef117f7-5cfa-4413-8c0c-1bb8a260749a","Type":"ContainerDied","Data":"b2a32d37bf7fbbb41f73f04a4a14e3e83d7edbd5cf5f620bb67aa141ea285f07"} Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.961744 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2a32d37bf7fbbb41f73f04a4a14e3e83d7edbd5cf5f620bb67aa141ea285f07" Dec 17 08:49:53 crc kubenswrapper[4966]: I1217 08:49:53.961727 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8zrh" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.047614 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4"] Dec 17 08:49:54 crc kubenswrapper[4966]: E1217 08:49:54.048352 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ef117f7-5cfa-4413-8c0c-1bb8a260749a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.048370 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ef117f7-5cfa-4413-8c0c-1bb8a260749a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 17 08:49:54 crc kubenswrapper[4966]: E1217 08:49:54.048407 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7dff045-586c-4bcd-a312-808d969ea508" containerName="extract-content" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.048414 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7dff045-586c-4bcd-a312-808d969ea508" containerName="extract-content" Dec 17 08:49:54 crc kubenswrapper[4966]: E1217 08:49:54.048424 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7dff045-586c-4bcd-a312-808d969ea508" containerName="registry-server" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.048430 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7dff045-586c-4bcd-a312-808d969ea508" containerName="registry-server" Dec 17 08:49:54 crc kubenswrapper[4966]: E1217 08:49:54.048440 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7dff045-586c-4bcd-a312-808d969ea508" containerName="extract-utilities" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.048447 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7dff045-586c-4bcd-a312-808d969ea508" containerName="extract-utilities" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.048632 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ef117f7-5cfa-4413-8c0c-1bb8a260749a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.048654 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7dff045-586c-4bcd-a312-808d969ea508" containerName="registry-server" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.049416 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.053692 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.053814 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.057364 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.059362 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.062544 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4"] Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.174445 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.174942 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rd9f\" (UniqueName: \"kubernetes.io/projected/c8f14153-71ae-460a-bb7b-61744fc40657-kube-api-access-8rd9f\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.175141 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.276855 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rd9f\" (UniqueName: \"kubernetes.io/projected/c8f14153-71ae-460a-bb7b-61744fc40657-kube-api-access-8rd9f\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.276952 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.276996 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.280417 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.281507 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.301534 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rd9f\" (UniqueName: \"kubernetes.io/projected/c8f14153-71ae-460a-bb7b-61744fc40657-kube-api-access-8rd9f\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.367308 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.945665 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4"] Dec 17 08:49:54 crc kubenswrapper[4966]: I1217 08:49:54.971606 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" event={"ID":"c8f14153-71ae-460a-bb7b-61744fc40657","Type":"ContainerStarted","Data":"64d0cce2d9429499844c0139a2fbfa7a5158587fa1435d6df4012c7b2ae51659"} Dec 17 08:49:55 crc kubenswrapper[4966]: I1217 08:49:55.045238 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6jj7z"] Dec 17 08:49:55 crc kubenswrapper[4966]: I1217 08:49:55.056574 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-gkjk7"] Dec 17 08:49:55 crc kubenswrapper[4966]: I1217 08:49:55.073915 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6jj7z"] Dec 17 08:49:55 crc kubenswrapper[4966]: I1217 08:49:55.083847 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-gkjk7"] Dec 17 08:49:55 crc kubenswrapper[4966]: I1217 08:49:55.986561 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" event={"ID":"c8f14153-71ae-460a-bb7b-61744fc40657","Type":"ContainerStarted","Data":"7e4843cd6f58ba2a17bff9b5f257f8527c1f2c40d1de22252d99ac62e17eb744"} Dec 17 08:49:56 crc kubenswrapper[4966]: I1217 08:49:56.010202 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" podStartSLOduration=1.510368359 podStartE2EDuration="2.010182931s" podCreationTimestamp="2025-12-17 08:49:54 +0000 UTC" firstStartedPulling="2025-12-17 08:49:54.954445353 +0000 UTC m=+1730.499515295" lastFinishedPulling="2025-12-17 08:49:55.454259915 +0000 UTC m=+1730.999329867" observedRunningTime="2025-12-17 08:49:56.003285733 +0000 UTC m=+1731.548355685" watchObservedRunningTime="2025-12-17 08:49:56.010182931 +0000 UTC m=+1731.555252903" Dec 17 08:49:56 crc kubenswrapper[4966]: I1217 08:49:56.847377 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95497f6e-9ff6-457b-a47a-3747cc4ddc69" path="/var/lib/kubelet/pods/95497f6e-9ff6-457b-a47a-3747cc4ddc69/volumes" Dec 17 08:49:56 crc kubenswrapper[4966]: I1217 08:49:56.849425 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eccb295b-14f1-43a9-98c8-673e2ce78add" path="/var/lib/kubelet/pods/eccb295b-14f1-43a9-98c8-673e2ce78add/volumes" Dec 17 08:49:57 crc kubenswrapper[4966]: I1217 08:49:57.830543 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:49:57 crc kubenswrapper[4966]: E1217 08:49:57.831594 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:50:06 crc kubenswrapper[4966]: I1217 08:50:06.042395 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9vf55"] Dec 17 08:50:06 crc kubenswrapper[4966]: I1217 08:50:06.051069 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9vf55"] Dec 17 08:50:06 crc kubenswrapper[4966]: I1217 08:50:06.843219 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ceb72da-ea40-4c24-8b26-94b10ba9cc6a" path="/var/lib/kubelet/pods/2ceb72da-ea40-4c24-8b26-94b10ba9cc6a/volumes" Dec 17 08:50:09 crc kubenswrapper[4966]: I1217 08:50:09.032306 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-rqvfb"] Dec 17 08:50:09 crc kubenswrapper[4966]: I1217 08:50:09.041052 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-9bd96"] Dec 17 08:50:09 crc kubenswrapper[4966]: I1217 08:50:09.049838 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-rqvfb"] Dec 17 08:50:09 crc kubenswrapper[4966]: I1217 08:50:09.066450 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-9bd96"] Dec 17 08:50:09 crc kubenswrapper[4966]: I1217 08:50:09.831972 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:50:09 crc kubenswrapper[4966]: E1217 08:50:09.832581 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:50:10 crc kubenswrapper[4966]: I1217 08:50:10.843060 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bca810d-b802-481a-8792-3be9b6a141bc" path="/var/lib/kubelet/pods/3bca810d-b802-481a-8792-3be9b6a141bc/volumes" Dec 17 08:50:10 crc kubenswrapper[4966]: I1217 08:50:10.844790 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7351028-bc40-4197-baa1-9fb0c525170d" path="/var/lib/kubelet/pods/b7351028-bc40-4197-baa1-9fb0c525170d/volumes" Dec 17 08:50:22 crc kubenswrapper[4966]: I1217 08:50:22.831105 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:50:22 crc kubenswrapper[4966]: E1217 08:50:22.832483 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:50:36 crc kubenswrapper[4966]: I1217 08:50:36.830848 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:50:36 crc kubenswrapper[4966]: E1217 08:50:36.831707 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:50:50 crc kubenswrapper[4966]: I1217 08:50:50.126154 4966 scope.go:117] "RemoveContainer" containerID="d6333c355cf08791bb1a8ac20b4ad7952b8619bcee1c6c44408c3f020e7e0342" Dec 17 08:50:50 crc kubenswrapper[4966]: I1217 08:50:50.160112 4966 scope.go:117] "RemoveContainer" containerID="5babfdce6e01bd62acd946a7fa8de1a13d7e35a1722e62ce0e5bde6558519fef" Dec 17 08:50:50 crc kubenswrapper[4966]: I1217 08:50:50.205340 4966 scope.go:117] "RemoveContainer" containerID="199905196c697327e46f3abf7b458fcc5d98f4c441c75c4866f5aadfc795f895" Dec 17 08:50:50 crc kubenswrapper[4966]: I1217 08:50:50.262021 4966 scope.go:117] "RemoveContainer" containerID="813da29eb9a6d3330fe50d3eef0a482abb8877d463e871b4ff9adaca45841ff3" Dec 17 08:50:50 crc kubenswrapper[4966]: I1217 08:50:50.294605 4966 scope.go:117] "RemoveContainer" containerID="5c7c6d66f7d0164b94f4acbc81db1caba0184245b63f6bb8e83580d4f8db3546" Dec 17 08:50:50 crc kubenswrapper[4966]: I1217 08:50:50.831044 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:50:50 crc kubenswrapper[4966]: E1217 08:50:50.831798 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.047463 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-67a5-account-create-update-qdhbr"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.055635 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-67a5-account-create-update-qdhbr"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.066054 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-c4llv"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.075521 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-7d41-account-create-update-x4fn4"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.084918 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-xlrkm"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.092109 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-c4llv"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.099091 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-spxpd"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.105547 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-e4a6-account-create-update-q8vbr"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.111863 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-7d41-account-create-update-x4fn4"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.119674 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-e4a6-account-create-update-q8vbr"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.126715 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-spxpd"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.133796 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-xlrkm"] Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.862045 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0644984a-4f00-4b7a-9a05-d9e43097a7c4" path="/var/lib/kubelet/pods/0644984a-4f00-4b7a-9a05-d9e43097a7c4/volumes" Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.863722 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b302878-13e7-4967-aaf1-58292a7237e5" path="/var/lib/kubelet/pods/2b302878-13e7-4967-aaf1-58292a7237e5/volumes" Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.865174 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b05f9d6-237c-4cb0-91b9-505715528c7d" path="/var/lib/kubelet/pods/6b05f9d6-237c-4cb0-91b9-505715528c7d/volumes" Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.866647 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9ea912d-6078-4ab1-8188-f1e2927eb12a" path="/var/lib/kubelet/pods/c9ea912d-6078-4ab1-8188-f1e2927eb12a/volumes" Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.868519 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce426531-9f5f-4c9d-8bea-e22d666fe95d" path="/var/lib/kubelet/pods/ce426531-9f5f-4c9d-8bea-e22d666fe95d/volumes" Dec 17 08:50:54 crc kubenswrapper[4966]: I1217 08:50:54.869826 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fed2ac76-ebb0-481f-82e1-55397f658c99" path="/var/lib/kubelet/pods/fed2ac76-ebb0-481f-82e1-55397f658c99/volumes" Dec 17 08:51:04 crc kubenswrapper[4966]: I1217 08:51:04.837273 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:51:04 crc kubenswrapper[4966]: E1217 08:51:04.838030 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:51:19 crc kubenswrapper[4966]: I1217 08:51:19.831095 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:51:19 crc kubenswrapper[4966]: E1217 08:51:19.831864 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:51:34 crc kubenswrapper[4966]: I1217 08:51:34.836668 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:51:34 crc kubenswrapper[4966]: E1217 08:51:34.837418 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:51:47 crc kubenswrapper[4966]: I1217 08:51:47.045079 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzwzg"] Dec 17 08:51:47 crc kubenswrapper[4966]: I1217 08:51:47.056772 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-hzwzg"] Dec 17 08:51:48 crc kubenswrapper[4966]: I1217 08:51:48.843448 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f6a226-d49b-4726-8116-4b9ab2837338" path="/var/lib/kubelet/pods/02f6a226-d49b-4726-8116-4b9ab2837338/volumes" Dec 17 08:51:49 crc kubenswrapper[4966]: I1217 08:51:49.831209 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:51:49 crc kubenswrapper[4966]: E1217 08:51:49.832245 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:51:50 crc kubenswrapper[4966]: I1217 08:51:50.424554 4966 scope.go:117] "RemoveContainer" containerID="af47b553eb89f736be25a50d8ab3e9a39064e822b0f742f58fa47098644aa92f" Dec 17 08:51:50 crc kubenswrapper[4966]: I1217 08:51:50.460292 4966 scope.go:117] "RemoveContainer" containerID="abe99bca543fdf2cf906ae9e9294ec48c8366d00058c0213a01458e6fb3c7664" Dec 17 08:51:50 crc kubenswrapper[4966]: I1217 08:51:50.517999 4966 scope.go:117] "RemoveContainer" containerID="c0754d5a0deea891d2ebd07e50c4cd12a376080dc1b6c8a5a2a5342077161320" Dec 17 08:51:50 crc kubenswrapper[4966]: I1217 08:51:50.554600 4966 scope.go:117] "RemoveContainer" containerID="c03aac27db194fa66d8f77291e0f1b16dda26943f297720177bd3bf48ed980a5" Dec 17 08:51:50 crc kubenswrapper[4966]: I1217 08:51:50.610187 4966 scope.go:117] "RemoveContainer" containerID="5ba9b49c17b666885899786efb3d50937ceb85900c2a06b6198056f893f3000e" Dec 17 08:51:50 crc kubenswrapper[4966]: I1217 08:51:50.657412 4966 scope.go:117] "RemoveContainer" containerID="ffad271bb68bd78bf50f98a591a054eecfb0aacaec261acef467f5c1dd4af4ca" Dec 17 08:51:50 crc kubenswrapper[4966]: I1217 08:51:50.710790 4966 scope.go:117] "RemoveContainer" containerID="21cb6a52d5dbe8f6273ea2d4a3330d298f871767fe3b4b5b28e6a463c09fef72" Dec 17 08:52:00 crc kubenswrapper[4966]: I1217 08:52:00.830788 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:52:00 crc kubenswrapper[4966]: E1217 08:52:00.831681 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:52:12 crc kubenswrapper[4966]: I1217 08:52:12.041683 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-7bzdn"] Dec 17 08:52:12 crc kubenswrapper[4966]: I1217 08:52:12.053571 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qws2d"] Dec 17 08:52:12 crc kubenswrapper[4966]: I1217 08:52:12.063803 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qws2d"] Dec 17 08:52:12 crc kubenswrapper[4966]: I1217 08:52:12.071965 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-7bzdn"] Dec 17 08:52:12 crc kubenswrapper[4966]: I1217 08:52:12.849339 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="072c5a1a-6292-4420-b5a1-dcd5dd93c5aa" path="/var/lib/kubelet/pods/072c5a1a-6292-4420-b5a1-dcd5dd93c5aa/volumes" Dec 17 08:52:12 crc kubenswrapper[4966]: I1217 08:52:12.851845 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="787797e9-40fd-4108-8448-65dd7ae41e2e" path="/var/lib/kubelet/pods/787797e9-40fd-4108-8448-65dd7ae41e2e/volumes" Dec 17 08:52:15 crc kubenswrapper[4966]: I1217 08:52:15.831360 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:52:15 crc kubenswrapper[4966]: E1217 08:52:15.833199 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:52:21 crc kubenswrapper[4966]: I1217 08:52:21.036470 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tdhl5"] Dec 17 08:52:21 crc kubenswrapper[4966]: I1217 08:52:21.045590 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tdhl5"] Dec 17 08:52:23 crc kubenswrapper[4966]: I1217 08:52:23.477152 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a272c7ab-dc3f-49af-9f02-6fe4b552c217" path="/var/lib/kubelet/pods/a272c7ab-dc3f-49af-9f02-6fe4b552c217/volumes" Dec 17 08:52:30 crc kubenswrapper[4966]: I1217 08:52:30.831251 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:52:30 crc kubenswrapper[4966]: E1217 08:52:30.832264 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:52:33 crc kubenswrapper[4966]: I1217 08:52:33.927864 4966 generic.go:334] "Generic (PLEG): container finished" podID="c8f14153-71ae-460a-bb7b-61744fc40657" containerID="7e4843cd6f58ba2a17bff9b5f257f8527c1f2c40d1de22252d99ac62e17eb744" exitCode=0 Dec 17 08:52:33 crc kubenswrapper[4966]: I1217 08:52:33.928099 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" event={"ID":"c8f14153-71ae-460a-bb7b-61744fc40657","Type":"ContainerDied","Data":"7e4843cd6f58ba2a17bff9b5f257f8527c1f2c40d1de22252d99ac62e17eb744"} Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.337807 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.488629 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-ssh-key\") pod \"c8f14153-71ae-460a-bb7b-61744fc40657\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.488861 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-inventory\") pod \"c8f14153-71ae-460a-bb7b-61744fc40657\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.488912 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rd9f\" (UniqueName: \"kubernetes.io/projected/c8f14153-71ae-460a-bb7b-61744fc40657-kube-api-access-8rd9f\") pod \"c8f14153-71ae-460a-bb7b-61744fc40657\" (UID: \"c8f14153-71ae-460a-bb7b-61744fc40657\") " Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.495460 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8f14153-71ae-460a-bb7b-61744fc40657-kube-api-access-8rd9f" (OuterVolumeSpecName: "kube-api-access-8rd9f") pod "c8f14153-71ae-460a-bb7b-61744fc40657" (UID: "c8f14153-71ae-460a-bb7b-61744fc40657"). InnerVolumeSpecName "kube-api-access-8rd9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.525580 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c8f14153-71ae-460a-bb7b-61744fc40657" (UID: "c8f14153-71ae-460a-bb7b-61744fc40657"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.529129 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-inventory" (OuterVolumeSpecName: "inventory") pod "c8f14153-71ae-460a-bb7b-61744fc40657" (UID: "c8f14153-71ae-460a-bb7b-61744fc40657"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.590611 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.590648 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rd9f\" (UniqueName: \"kubernetes.io/projected/c8f14153-71ae-460a-bb7b-61744fc40657-kube-api-access-8rd9f\") on node \"crc\" DevicePath \"\"" Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.590668 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c8f14153-71ae-460a-bb7b-61744fc40657-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.951095 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" event={"ID":"c8f14153-71ae-460a-bb7b-61744fc40657","Type":"ContainerDied","Data":"64d0cce2d9429499844c0139a2fbfa7a5158587fa1435d6df4012c7b2ae51659"} Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.951133 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64d0cce2d9429499844c0139a2fbfa7a5158587fa1435d6df4012c7b2ae51659" Dec 17 08:52:35 crc kubenswrapper[4966]: I1217 08:52:35.951220 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fvkk4" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.077800 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx"] Dec 17 08:52:36 crc kubenswrapper[4966]: E1217 08:52:36.078639 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8f14153-71ae-460a-bb7b-61744fc40657" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.078661 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8f14153-71ae-460a-bb7b-61744fc40657" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.079359 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8f14153-71ae-460a-bb7b-61744fc40657" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.082801 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.087098 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.087157 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.087294 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.087294 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.108457 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx"] Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.201286 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.201369 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.201426 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjg96\" (UniqueName: \"kubernetes.io/projected/dd4cb85b-21e6-4798-b0c8-45b79d946037-kube-api-access-wjg96\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.303930 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.304034 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjg96\" (UniqueName: \"kubernetes.io/projected/dd4cb85b-21e6-4798-b0c8-45b79d946037-kube-api-access-wjg96\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.304210 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.310003 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.312323 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.320407 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjg96\" (UniqueName: \"kubernetes.io/projected/dd4cb85b-21e6-4798-b0c8-45b79d946037-kube-api-access-wjg96\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.410261 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.974302 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx"] Dec 17 08:52:36 crc kubenswrapper[4966]: I1217 08:52:36.998329 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 08:52:37 crc kubenswrapper[4966]: I1217 08:52:37.976395 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" event={"ID":"dd4cb85b-21e6-4798-b0c8-45b79d946037","Type":"ContainerStarted","Data":"b2f1d57b0ea34ab97639396b3be6f5e71495dba71a0639ea67f95206e1c9b07f"} Dec 17 08:52:38 crc kubenswrapper[4966]: I1217 08:52:38.987126 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" event={"ID":"dd4cb85b-21e6-4798-b0c8-45b79d946037","Type":"ContainerStarted","Data":"55a91a0a8129eb6fc20e871e118ee44bdbb789f492f5f9fcb877bd7a4e8ecfab"} Dec 17 08:52:39 crc kubenswrapper[4966]: I1217 08:52:39.009578 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" podStartSLOduration=1.359147316 podStartE2EDuration="3.009554393s" podCreationTimestamp="2025-12-17 08:52:36 +0000 UTC" firstStartedPulling="2025-12-17 08:52:36.99813369 +0000 UTC m=+1892.543203622" lastFinishedPulling="2025-12-17 08:52:38.648540727 +0000 UTC m=+1894.193610699" observedRunningTime="2025-12-17 08:52:39.002983183 +0000 UTC m=+1894.548053135" watchObservedRunningTime="2025-12-17 08:52:39.009554393 +0000 UTC m=+1894.554624335" Dec 17 08:52:42 crc kubenswrapper[4966]: I1217 08:52:42.830390 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:52:42 crc kubenswrapper[4966]: E1217 08:52:42.830889 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:52:50 crc kubenswrapper[4966]: I1217 08:52:50.840308 4966 scope.go:117] "RemoveContainer" containerID="069702d674a0dfc7ece303cf410082fe66470f695b0900569072234a90df79f9" Dec 17 08:52:50 crc kubenswrapper[4966]: I1217 08:52:50.870922 4966 scope.go:117] "RemoveContainer" containerID="4eb94a4fb5c18d6acd037c4294a3352d575e3b79ac2e5f740d46fa74aa105530" Dec 17 08:52:50 crc kubenswrapper[4966]: I1217 08:52:50.932387 4966 scope.go:117] "RemoveContainer" containerID="e4ef0f804df1958cad65b3e71ba5895646044cf889412cd1913439c97da4c73f" Dec 17 08:52:53 crc kubenswrapper[4966]: I1217 08:52:53.831082 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:52:53 crc kubenswrapper[4966]: E1217 08:52:53.831827 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:52:58 crc kubenswrapper[4966]: I1217 08:52:58.042251 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-x4f2k"] Dec 17 08:52:58 crc kubenswrapper[4966]: I1217 08:52:58.051436 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-x4f2k"] Dec 17 08:52:58 crc kubenswrapper[4966]: I1217 08:52:58.841471 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c08b24e-7324-49f1-b084-fd2f866dd7c4" path="/var/lib/kubelet/pods/4c08b24e-7324-49f1-b084-fd2f866dd7c4/volumes" Dec 17 08:53:07 crc kubenswrapper[4966]: I1217 08:53:07.830569 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:53:07 crc kubenswrapper[4966]: E1217 08:53:07.836683 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:53:20 crc kubenswrapper[4966]: I1217 08:53:20.831015 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:53:20 crc kubenswrapper[4966]: E1217 08:53:20.831749 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:53:34 crc kubenswrapper[4966]: I1217 08:53:34.840127 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:53:34 crc kubenswrapper[4966]: E1217 08:53:34.841111 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:53:47 crc kubenswrapper[4966]: I1217 08:53:47.830489 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:53:47 crc kubenswrapper[4966]: E1217 08:53:47.831293 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:53:51 crc kubenswrapper[4966]: I1217 08:53:51.050029 4966 scope.go:117] "RemoveContainer" containerID="ef731192a3d3c582c54662603c920c08d8faf1d504da7a9e068e49425bbaa852" Dec 17 08:54:02 crc kubenswrapper[4966]: I1217 08:54:02.830663 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:54:02 crc kubenswrapper[4966]: E1217 08:54:02.831793 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:54:12 crc kubenswrapper[4966]: I1217 08:54:12.791681 4966 generic.go:334] "Generic (PLEG): container finished" podID="dd4cb85b-21e6-4798-b0c8-45b79d946037" containerID="55a91a0a8129eb6fc20e871e118ee44bdbb789f492f5f9fcb877bd7a4e8ecfab" exitCode=0 Dec 17 08:54:12 crc kubenswrapper[4966]: I1217 08:54:12.791773 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" event={"ID":"dd4cb85b-21e6-4798-b0c8-45b79d946037","Type":"ContainerDied","Data":"55a91a0a8129eb6fc20e871e118ee44bdbb789f492f5f9fcb877bd7a4e8ecfab"} Dec 17 08:54:13 crc kubenswrapper[4966]: I1217 08:54:13.831189 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:54:13 crc kubenswrapper[4966]: E1217 08:54:13.831490 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.213236 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.251637 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-inventory\") pod \"dd4cb85b-21e6-4798-b0c8-45b79d946037\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.251719 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjg96\" (UniqueName: \"kubernetes.io/projected/dd4cb85b-21e6-4798-b0c8-45b79d946037-kube-api-access-wjg96\") pod \"dd4cb85b-21e6-4798-b0c8-45b79d946037\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.251959 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-ssh-key\") pod \"dd4cb85b-21e6-4798-b0c8-45b79d946037\" (UID: \"dd4cb85b-21e6-4798-b0c8-45b79d946037\") " Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.275560 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd4cb85b-21e6-4798-b0c8-45b79d946037-kube-api-access-wjg96" (OuterVolumeSpecName: "kube-api-access-wjg96") pod "dd4cb85b-21e6-4798-b0c8-45b79d946037" (UID: "dd4cb85b-21e6-4798-b0c8-45b79d946037"). InnerVolumeSpecName "kube-api-access-wjg96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.288706 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dd4cb85b-21e6-4798-b0c8-45b79d946037" (UID: "dd4cb85b-21e6-4798-b0c8-45b79d946037"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.297099 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-inventory" (OuterVolumeSpecName: "inventory") pod "dd4cb85b-21e6-4798-b0c8-45b79d946037" (UID: "dd4cb85b-21e6-4798-b0c8-45b79d946037"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.354343 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.354376 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd4cb85b-21e6-4798-b0c8-45b79d946037-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.354387 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjg96\" (UniqueName: \"kubernetes.io/projected/dd4cb85b-21e6-4798-b0c8-45b79d946037-kube-api-access-wjg96\") on node \"crc\" DevicePath \"\"" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.817541 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" event={"ID":"dd4cb85b-21e6-4798-b0c8-45b79d946037","Type":"ContainerDied","Data":"b2f1d57b0ea34ab97639396b3be6f5e71495dba71a0639ea67f95206e1c9b07f"} Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.817896 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2f1d57b0ea34ab97639396b3be6f5e71495dba71a0639ea67f95206e1c9b07f" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.817967 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-7nbxx" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.926044 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8"] Dec 17 08:54:14 crc kubenswrapper[4966]: E1217 08:54:14.926507 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd4cb85b-21e6-4798-b0c8-45b79d946037" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.926519 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd4cb85b-21e6-4798-b0c8-45b79d946037" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.926694 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd4cb85b-21e6-4798-b0c8-45b79d946037" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.927990 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.930298 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.933032 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.933420 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.934936 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.942152 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8"] Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.965066 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.965134 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:14 crc kubenswrapper[4966]: I1217 08:54:14.965347 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvfv8\" (UniqueName: \"kubernetes.io/projected/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-kube-api-access-mvfv8\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:15 crc kubenswrapper[4966]: I1217 08:54:15.068033 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:15 crc kubenswrapper[4966]: I1217 08:54:15.068086 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:15 crc kubenswrapper[4966]: I1217 08:54:15.068200 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvfv8\" (UniqueName: \"kubernetes.io/projected/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-kube-api-access-mvfv8\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:15 crc kubenswrapper[4966]: I1217 08:54:15.075593 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:15 crc kubenswrapper[4966]: I1217 08:54:15.080404 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:15 crc kubenswrapper[4966]: I1217 08:54:15.085565 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvfv8\" (UniqueName: \"kubernetes.io/projected/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-kube-api-access-mvfv8\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:15 crc kubenswrapper[4966]: I1217 08:54:15.244677 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:15 crc kubenswrapper[4966]: W1217 08:54:15.826264 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd9f447e_71f2_48e9_ab4b_7b706d40c2a7.slice/crio-6072f78094831d703b36980bad96bd3e77b6103e1a27f7f7cc31f2e5ee9a49e0 WatchSource:0}: Error finding container 6072f78094831d703b36980bad96bd3e77b6103e1a27f7f7cc31f2e5ee9a49e0: Status 404 returned error can't find the container with id 6072f78094831d703b36980bad96bd3e77b6103e1a27f7f7cc31f2e5ee9a49e0 Dec 17 08:54:15 crc kubenswrapper[4966]: I1217 08:54:15.837650 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8"] Dec 17 08:54:16 crc kubenswrapper[4966]: I1217 08:54:16.982118 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" event={"ID":"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7","Type":"ContainerStarted","Data":"30baa4008f783266ac3882f7cebcde5f4cf1cfcf8f2c6818fccc8a467d067f01"} Dec 17 08:54:16 crc kubenswrapper[4966]: I1217 08:54:16.983641 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" event={"ID":"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7","Type":"ContainerStarted","Data":"6072f78094831d703b36980bad96bd3e77b6103e1a27f7f7cc31f2e5ee9a49e0"} Dec 17 08:54:17 crc kubenswrapper[4966]: I1217 08:54:17.008614 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" podStartSLOduration=2.431742988 podStartE2EDuration="3.008595896s" podCreationTimestamp="2025-12-17 08:54:14 +0000 UTC" firstStartedPulling="2025-12-17 08:54:15.828481288 +0000 UTC m=+1991.373551230" lastFinishedPulling="2025-12-17 08:54:16.405334196 +0000 UTC m=+1991.950404138" observedRunningTime="2025-12-17 08:54:17.006323883 +0000 UTC m=+1992.551393825" watchObservedRunningTime="2025-12-17 08:54:17.008595896 +0000 UTC m=+1992.553665838" Dec 17 08:54:23 crc kubenswrapper[4966]: I1217 08:54:23.028608 4966 generic.go:334] "Generic (PLEG): container finished" podID="bd9f447e-71f2-48e9-ab4b-7b706d40c2a7" containerID="30baa4008f783266ac3882f7cebcde5f4cf1cfcf8f2c6818fccc8a467d067f01" exitCode=0 Dec 17 08:54:23 crc kubenswrapper[4966]: I1217 08:54:23.028652 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" event={"ID":"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7","Type":"ContainerDied","Data":"30baa4008f783266ac3882f7cebcde5f4cf1cfcf8f2c6818fccc8a467d067f01"} Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.455666 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.563638 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvfv8\" (UniqueName: \"kubernetes.io/projected/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-kube-api-access-mvfv8\") pod \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.563739 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-inventory\") pod \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.563764 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-ssh-key\") pod \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\" (UID: \"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7\") " Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.569363 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-kube-api-access-mvfv8" (OuterVolumeSpecName: "kube-api-access-mvfv8") pod "bd9f447e-71f2-48e9-ab4b-7b706d40c2a7" (UID: "bd9f447e-71f2-48e9-ab4b-7b706d40c2a7"). InnerVolumeSpecName "kube-api-access-mvfv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.595845 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bd9f447e-71f2-48e9-ab4b-7b706d40c2a7" (UID: "bd9f447e-71f2-48e9-ab4b-7b706d40c2a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.609154 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-inventory" (OuterVolumeSpecName: "inventory") pod "bd9f447e-71f2-48e9-ab4b-7b706d40c2a7" (UID: "bd9f447e-71f2-48e9-ab4b-7b706d40c2a7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.666052 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvfv8\" (UniqueName: \"kubernetes.io/projected/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-kube-api-access-mvfv8\") on node \"crc\" DevicePath \"\"" Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.666896 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:54:24 crc kubenswrapper[4966]: I1217 08:54:24.667022 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd9f447e-71f2-48e9-ab4b-7b706d40c2a7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.046086 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" event={"ID":"bd9f447e-71f2-48e9-ab4b-7b706d40c2a7","Type":"ContainerDied","Data":"6072f78094831d703b36980bad96bd3e77b6103e1a27f7f7cc31f2e5ee9a49e0"} Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.046128 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6072f78094831d703b36980bad96bd3e77b6103e1a27f7f7cc31f2e5ee9a49e0" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.046184 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kxgd8" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.129402 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx"] Dec 17 08:54:25 crc kubenswrapper[4966]: E1217 08:54:25.129797 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9f447e-71f2-48e9-ab4b-7b706d40c2a7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.129814 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9f447e-71f2-48e9-ab4b-7b706d40c2a7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.130013 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd9f447e-71f2-48e9-ab4b-7b706d40c2a7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.130912 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.137629 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.137989 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.138017 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.138124 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.148011 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx"] Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.278602 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-448fx\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.278655 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-448fx\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.278741 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pskzf\" (UniqueName: \"kubernetes.io/projected/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-kube-api-access-pskzf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-448fx\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.380036 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pskzf\" (UniqueName: \"kubernetes.io/projected/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-kube-api-access-pskzf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-448fx\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.380380 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-448fx\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.380423 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-448fx\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.384527 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-448fx\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.384528 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-448fx\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.399276 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pskzf\" (UniqueName: \"kubernetes.io/projected/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-kube-api-access-pskzf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-448fx\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:25 crc kubenswrapper[4966]: I1217 08:54:25.451921 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:54:26 crc kubenswrapper[4966]: I1217 08:54:26.027552 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx"] Dec 17 08:54:26 crc kubenswrapper[4966]: I1217 08:54:26.061453 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" event={"ID":"8ce543ee-2d25-4abe-9b0d-cdfe480339e5","Type":"ContainerStarted","Data":"1e74dd308eac925afbeed1e78de013fdc96f3f56341e70326af6376955ce8aa8"} Dec 17 08:54:27 crc kubenswrapper[4966]: I1217 08:54:27.069137 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" event={"ID":"8ce543ee-2d25-4abe-9b0d-cdfe480339e5","Type":"ContainerStarted","Data":"ab666fc74571861cfd37f42c8d980eff8e7480061534f7dc90d6f65528e812ac"} Dec 17 08:54:27 crc kubenswrapper[4966]: I1217 08:54:27.086128 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" podStartSLOduration=1.454606816 podStartE2EDuration="2.086107776s" podCreationTimestamp="2025-12-17 08:54:25 +0000 UTC" firstStartedPulling="2025-12-17 08:54:26.036103281 +0000 UTC m=+2001.581173213" lastFinishedPulling="2025-12-17 08:54:26.667604231 +0000 UTC m=+2002.212674173" observedRunningTime="2025-12-17 08:54:27.08255359 +0000 UTC m=+2002.627623542" watchObservedRunningTime="2025-12-17 08:54:27.086107776 +0000 UTC m=+2002.631177718" Dec 17 08:54:28 crc kubenswrapper[4966]: I1217 08:54:28.831845 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:54:29 crc kubenswrapper[4966]: I1217 08:54:29.087367 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"4757810d46cf706e3c032dafdc0e43a6e21fb930b0834212015c013294922d1f"} Dec 17 08:55:08 crc kubenswrapper[4966]: I1217 08:55:08.485451 4966 generic.go:334] "Generic (PLEG): container finished" podID="8ce543ee-2d25-4abe-9b0d-cdfe480339e5" containerID="ab666fc74571861cfd37f42c8d980eff8e7480061534f7dc90d6f65528e812ac" exitCode=0 Dec 17 08:55:08 crc kubenswrapper[4966]: I1217 08:55:08.485538 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" event={"ID":"8ce543ee-2d25-4abe-9b0d-cdfe480339e5","Type":"ContainerDied","Data":"ab666fc74571861cfd37f42c8d980eff8e7480061534f7dc90d6f65528e812ac"} Dec 17 08:55:09 crc kubenswrapper[4966]: I1217 08:55:09.972600 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.129753 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-ssh-key\") pod \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.130221 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pskzf\" (UniqueName: \"kubernetes.io/projected/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-kube-api-access-pskzf\") pod \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.130417 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-inventory\") pod \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\" (UID: \"8ce543ee-2d25-4abe-9b0d-cdfe480339e5\") " Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.143068 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-kube-api-access-pskzf" (OuterVolumeSpecName: "kube-api-access-pskzf") pod "8ce543ee-2d25-4abe-9b0d-cdfe480339e5" (UID: "8ce543ee-2d25-4abe-9b0d-cdfe480339e5"). InnerVolumeSpecName "kube-api-access-pskzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.157628 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-inventory" (OuterVolumeSpecName: "inventory") pod "8ce543ee-2d25-4abe-9b0d-cdfe480339e5" (UID: "8ce543ee-2d25-4abe-9b0d-cdfe480339e5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.163100 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8ce543ee-2d25-4abe-9b0d-cdfe480339e5" (UID: "8ce543ee-2d25-4abe-9b0d-cdfe480339e5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.233028 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pskzf\" (UniqueName: \"kubernetes.io/projected/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-kube-api-access-pskzf\") on node \"crc\" DevicePath \"\"" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.233091 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.233110 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ce543ee-2d25-4abe-9b0d-cdfe480339e5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.505044 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" event={"ID":"8ce543ee-2d25-4abe-9b0d-cdfe480339e5","Type":"ContainerDied","Data":"1e74dd308eac925afbeed1e78de013fdc96f3f56341e70326af6376955ce8aa8"} Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.505083 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e74dd308eac925afbeed1e78de013fdc96f3f56341e70326af6376955ce8aa8" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.505139 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-448fx" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.613185 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg"] Dec 17 08:55:10 crc kubenswrapper[4966]: E1217 08:55:10.614489 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce543ee-2d25-4abe-9b0d-cdfe480339e5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.614508 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce543ee-2d25-4abe-9b0d-cdfe480339e5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.614704 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ce543ee-2d25-4abe-9b0d-cdfe480339e5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.615361 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.618312 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.618675 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.620428 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.622569 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.635486 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg"] Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.744976 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd9c4\" (UniqueName: \"kubernetes.io/projected/d5361b50-b76b-4903-ab66-57edeab3004d-kube-api-access-rd9c4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f29xg\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.745044 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f29xg\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.745191 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f29xg\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.846389 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd9c4\" (UniqueName: \"kubernetes.io/projected/d5361b50-b76b-4903-ab66-57edeab3004d-kube-api-access-rd9c4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f29xg\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.846443 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f29xg\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.846529 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f29xg\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.850515 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f29xg\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.850573 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f29xg\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.870257 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd9c4\" (UniqueName: \"kubernetes.io/projected/d5361b50-b76b-4903-ab66-57edeab3004d-kube-api-access-rd9c4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f29xg\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:10 crc kubenswrapper[4966]: I1217 08:55:10.949027 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:55:11 crc kubenswrapper[4966]: I1217 08:55:11.493607 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg"] Dec 17 08:55:11 crc kubenswrapper[4966]: I1217 08:55:11.515226 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" event={"ID":"d5361b50-b76b-4903-ab66-57edeab3004d","Type":"ContainerStarted","Data":"952a6e8d8a297ebd1a7e0785e991b73d495b311491d6ed6933cbeea99444aece"} Dec 17 08:55:12 crc kubenswrapper[4966]: I1217 08:55:12.526282 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" event={"ID":"d5361b50-b76b-4903-ab66-57edeab3004d","Type":"ContainerStarted","Data":"4ea02f193264aa1a0075386d968d3adb33541939113d6aeedd1150f214daa105"} Dec 17 08:55:12 crc kubenswrapper[4966]: I1217 08:55:12.549508 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" podStartSLOduration=2.025101975 podStartE2EDuration="2.54948325s" podCreationTimestamp="2025-12-17 08:55:10 +0000 UTC" firstStartedPulling="2025-12-17 08:55:11.505255192 +0000 UTC m=+2047.050325134" lastFinishedPulling="2025-12-17 08:55:12.029636467 +0000 UTC m=+2047.574706409" observedRunningTime="2025-12-17 08:55:12.541134752 +0000 UTC m=+2048.086204724" watchObservedRunningTime="2025-12-17 08:55:12.54948325 +0000 UTC m=+2048.094553202" Dec 17 08:56:08 crc kubenswrapper[4966]: I1217 08:56:08.059391 4966 generic.go:334] "Generic (PLEG): container finished" podID="d5361b50-b76b-4903-ab66-57edeab3004d" containerID="4ea02f193264aa1a0075386d968d3adb33541939113d6aeedd1150f214daa105" exitCode=0 Dec 17 08:56:08 crc kubenswrapper[4966]: I1217 08:56:08.059520 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" event={"ID":"d5361b50-b76b-4903-ab66-57edeab3004d","Type":"ContainerDied","Data":"4ea02f193264aa1a0075386d968d3adb33541939113d6aeedd1150f214daa105"} Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.491443 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.686531 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd9c4\" (UniqueName: \"kubernetes.io/projected/d5361b50-b76b-4903-ab66-57edeab3004d-kube-api-access-rd9c4\") pod \"d5361b50-b76b-4903-ab66-57edeab3004d\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.686770 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-ssh-key\") pod \"d5361b50-b76b-4903-ab66-57edeab3004d\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.686986 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-inventory\") pod \"d5361b50-b76b-4903-ab66-57edeab3004d\" (UID: \"d5361b50-b76b-4903-ab66-57edeab3004d\") " Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.692143 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5361b50-b76b-4903-ab66-57edeab3004d-kube-api-access-rd9c4" (OuterVolumeSpecName: "kube-api-access-rd9c4") pod "d5361b50-b76b-4903-ab66-57edeab3004d" (UID: "d5361b50-b76b-4903-ab66-57edeab3004d"). InnerVolumeSpecName "kube-api-access-rd9c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.717702 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d5361b50-b76b-4903-ab66-57edeab3004d" (UID: "d5361b50-b76b-4903-ab66-57edeab3004d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.719500 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-inventory" (OuterVolumeSpecName: "inventory") pod "d5361b50-b76b-4903-ab66-57edeab3004d" (UID: "d5361b50-b76b-4903-ab66-57edeab3004d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.789510 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd9c4\" (UniqueName: \"kubernetes.io/projected/d5361b50-b76b-4903-ab66-57edeab3004d-kube-api-access-rd9c4\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.789550 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:09 crc kubenswrapper[4966]: I1217 08:56:09.789562 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5361b50-b76b-4903-ab66-57edeab3004d-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.083412 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" event={"ID":"d5361b50-b76b-4903-ab66-57edeab3004d","Type":"ContainerDied","Data":"952a6e8d8a297ebd1a7e0785e991b73d495b311491d6ed6933cbeea99444aece"} Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.083456 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="952a6e8d8a297ebd1a7e0785e991b73d495b311491d6ed6933cbeea99444aece" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.083460 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f29xg" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.180475 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-tsmck"] Dec 17 08:56:10 crc kubenswrapper[4966]: E1217 08:56:10.180965 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5361b50-b76b-4903-ab66-57edeab3004d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.180990 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5361b50-b76b-4903-ab66-57edeab3004d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.181247 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5361b50-b76b-4903-ab66-57edeab3004d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.182028 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.184503 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.184856 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.194962 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.195240 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.196828 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-tsmck"] Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.298816 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdnhk\" (UniqueName: \"kubernetes.io/projected/aee9703c-279e-4ac4-8451-78b16be0b1d9-kube-api-access-qdnhk\") pod \"ssh-known-hosts-edpm-deployment-tsmck\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.298961 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-tsmck\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.299253 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-tsmck\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.401147 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-tsmck\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.401323 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdnhk\" (UniqueName: \"kubernetes.io/projected/aee9703c-279e-4ac4-8451-78b16be0b1d9-kube-api-access-qdnhk\") pod \"ssh-known-hosts-edpm-deployment-tsmck\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.401390 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-tsmck\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.405211 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-tsmck\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.405515 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-tsmck\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.424486 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdnhk\" (UniqueName: \"kubernetes.io/projected/aee9703c-279e-4ac4-8451-78b16be0b1d9-kube-api-access-qdnhk\") pod \"ssh-known-hosts-edpm-deployment-tsmck\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:10 crc kubenswrapper[4966]: I1217 08:56:10.551577 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:11 crc kubenswrapper[4966]: I1217 08:56:11.082411 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-tsmck"] Dec 17 08:56:12 crc kubenswrapper[4966]: I1217 08:56:12.121602 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" event={"ID":"aee9703c-279e-4ac4-8451-78b16be0b1d9","Type":"ContainerStarted","Data":"e8ef8bcffb196a14aa30383985c35fac0159e6e44f9f163e579c19dc091c08f5"} Dec 17 08:56:12 crc kubenswrapper[4966]: I1217 08:56:12.123779 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" event={"ID":"aee9703c-279e-4ac4-8451-78b16be0b1d9","Type":"ContainerStarted","Data":"5429f65f90544f4f48da846bd1bbef5b7636eb6df0393b58becc118e410e2dbb"} Dec 17 08:56:20 crc kubenswrapper[4966]: I1217 08:56:20.192140 4966 generic.go:334] "Generic (PLEG): container finished" podID="aee9703c-279e-4ac4-8451-78b16be0b1d9" containerID="e8ef8bcffb196a14aa30383985c35fac0159e6e44f9f163e579c19dc091c08f5" exitCode=0 Dec 17 08:56:20 crc kubenswrapper[4966]: I1217 08:56:20.192255 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" event={"ID":"aee9703c-279e-4ac4-8451-78b16be0b1d9","Type":"ContainerDied","Data":"e8ef8bcffb196a14aa30383985c35fac0159e6e44f9f163e579c19dc091c08f5"} Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.654758 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.781604 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-inventory-0\") pod \"aee9703c-279e-4ac4-8451-78b16be0b1d9\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.781724 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-ssh-key-openstack-edpm-ipam\") pod \"aee9703c-279e-4ac4-8451-78b16be0b1d9\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.782007 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdnhk\" (UniqueName: \"kubernetes.io/projected/aee9703c-279e-4ac4-8451-78b16be0b1d9-kube-api-access-qdnhk\") pod \"aee9703c-279e-4ac4-8451-78b16be0b1d9\" (UID: \"aee9703c-279e-4ac4-8451-78b16be0b1d9\") " Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.792277 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aee9703c-279e-4ac4-8451-78b16be0b1d9-kube-api-access-qdnhk" (OuterVolumeSpecName: "kube-api-access-qdnhk") pod "aee9703c-279e-4ac4-8451-78b16be0b1d9" (UID: "aee9703c-279e-4ac4-8451-78b16be0b1d9"). InnerVolumeSpecName "kube-api-access-qdnhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.808252 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "aee9703c-279e-4ac4-8451-78b16be0b1d9" (UID: "aee9703c-279e-4ac4-8451-78b16be0b1d9"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.810454 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "aee9703c-279e-4ac4-8451-78b16be0b1d9" (UID: "aee9703c-279e-4ac4-8451-78b16be0b1d9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.884153 4966 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.884190 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/aee9703c-279e-4ac4-8451-78b16be0b1d9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:21 crc kubenswrapper[4966]: I1217 08:56:21.884203 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdnhk\" (UniqueName: \"kubernetes.io/projected/aee9703c-279e-4ac4-8451-78b16be0b1d9-kube-api-access-qdnhk\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.215203 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" event={"ID":"aee9703c-279e-4ac4-8451-78b16be0b1d9","Type":"ContainerDied","Data":"5429f65f90544f4f48da846bd1bbef5b7636eb6df0393b58becc118e410e2dbb"} Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.215246 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5429f65f90544f4f48da846bd1bbef5b7636eb6df0393b58becc118e410e2dbb" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.215307 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-tsmck" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.293982 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk"] Dec 17 08:56:22 crc kubenswrapper[4966]: E1217 08:56:22.294619 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee9703c-279e-4ac4-8451-78b16be0b1d9" containerName="ssh-known-hosts-edpm-deployment" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.294641 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee9703c-279e-4ac4-8451-78b16be0b1d9" containerName="ssh-known-hosts-edpm-deployment" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.295001 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee9703c-279e-4ac4-8451-78b16be0b1d9" containerName="ssh-known-hosts-edpm-deployment" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.295743 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.303451 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.305235 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.305363 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.308050 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk"] Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.310210 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.394655 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6ffg\" (UniqueName: \"kubernetes.io/projected/4f164e75-9cf1-4cde-89ba-32a85cced4b5-kube-api-access-c6ffg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m6nlk\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.394974 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m6nlk\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.395142 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m6nlk\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.496241 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6ffg\" (UniqueName: \"kubernetes.io/projected/4f164e75-9cf1-4cde-89ba-32a85cced4b5-kube-api-access-c6ffg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m6nlk\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.496374 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m6nlk\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.496434 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m6nlk\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.501262 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m6nlk\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.502390 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m6nlk\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.516548 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6ffg\" (UniqueName: \"kubernetes.io/projected/4f164e75-9cf1-4cde-89ba-32a85cced4b5-kube-api-access-c6ffg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-m6nlk\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:22 crc kubenswrapper[4966]: I1217 08:56:22.665355 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:23 crc kubenswrapper[4966]: I1217 08:56:23.202267 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk"] Dec 17 08:56:23 crc kubenswrapper[4966]: W1217 08:56:23.210299 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f164e75_9cf1_4cde_89ba_32a85cced4b5.slice/crio-a6ee199c556f0d60414fe39941a08d4cb5a163dbbd5a974f7dc8031009e634be WatchSource:0}: Error finding container a6ee199c556f0d60414fe39941a08d4cb5a163dbbd5a974f7dc8031009e634be: Status 404 returned error can't find the container with id a6ee199c556f0d60414fe39941a08d4cb5a163dbbd5a974f7dc8031009e634be Dec 17 08:56:23 crc kubenswrapper[4966]: I1217 08:56:23.225603 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" event={"ID":"4f164e75-9cf1-4cde-89ba-32a85cced4b5","Type":"ContainerStarted","Data":"a6ee199c556f0d60414fe39941a08d4cb5a163dbbd5a974f7dc8031009e634be"} Dec 17 08:56:24 crc kubenswrapper[4966]: I1217 08:56:24.237519 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" event={"ID":"4f164e75-9cf1-4cde-89ba-32a85cced4b5","Type":"ContainerStarted","Data":"61f1298bd806b5a2f0290cc3551386e891e1d88dd0db4bdf8b032f69e92c497f"} Dec 17 08:56:24 crc kubenswrapper[4966]: I1217 08:56:24.257576 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" podStartSLOduration=1.766660415 podStartE2EDuration="2.257550126s" podCreationTimestamp="2025-12-17 08:56:22 +0000 UTC" firstStartedPulling="2025-12-17 08:56:23.212005183 +0000 UTC m=+2118.757075125" lastFinishedPulling="2025-12-17 08:56:23.702894894 +0000 UTC m=+2119.247964836" observedRunningTime="2025-12-17 08:56:24.255936822 +0000 UTC m=+2119.801006764" watchObservedRunningTime="2025-12-17 08:56:24.257550126 +0000 UTC m=+2119.802620078" Dec 17 08:56:33 crc kubenswrapper[4966]: I1217 08:56:33.331976 4966 generic.go:334] "Generic (PLEG): container finished" podID="4f164e75-9cf1-4cde-89ba-32a85cced4b5" containerID="61f1298bd806b5a2f0290cc3551386e891e1d88dd0db4bdf8b032f69e92c497f" exitCode=0 Dec 17 08:56:33 crc kubenswrapper[4966]: I1217 08:56:33.332176 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" event={"ID":"4f164e75-9cf1-4cde-89ba-32a85cced4b5","Type":"ContainerDied","Data":"61f1298bd806b5a2f0290cc3551386e891e1d88dd0db4bdf8b032f69e92c497f"} Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.821574 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.888326 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-ssh-key\") pod \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.888386 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-inventory\") pod \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.888443 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6ffg\" (UniqueName: \"kubernetes.io/projected/4f164e75-9cf1-4cde-89ba-32a85cced4b5-kube-api-access-c6ffg\") pod \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\" (UID: \"4f164e75-9cf1-4cde-89ba-32a85cced4b5\") " Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.909416 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f164e75-9cf1-4cde-89ba-32a85cced4b5-kube-api-access-c6ffg" (OuterVolumeSpecName: "kube-api-access-c6ffg") pod "4f164e75-9cf1-4cde-89ba-32a85cced4b5" (UID: "4f164e75-9cf1-4cde-89ba-32a85cced4b5"). InnerVolumeSpecName "kube-api-access-c6ffg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.917112 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4f164e75-9cf1-4cde-89ba-32a85cced4b5" (UID: "4f164e75-9cf1-4cde-89ba-32a85cced4b5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.922029 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-inventory" (OuterVolumeSpecName: "inventory") pod "4f164e75-9cf1-4cde-89ba-32a85cced4b5" (UID: "4f164e75-9cf1-4cde-89ba-32a85cced4b5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.990567 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.990596 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f164e75-9cf1-4cde-89ba-32a85cced4b5-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:34 crc kubenswrapper[4966]: I1217 08:56:34.990606 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6ffg\" (UniqueName: \"kubernetes.io/projected/4f164e75-9cf1-4cde-89ba-32a85cced4b5-kube-api-access-c6ffg\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.352418 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" event={"ID":"4f164e75-9cf1-4cde-89ba-32a85cced4b5","Type":"ContainerDied","Data":"a6ee199c556f0d60414fe39941a08d4cb5a163dbbd5a974f7dc8031009e634be"} Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.352720 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6ee199c556f0d60414fe39941a08d4cb5a163dbbd5a974f7dc8031009e634be" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.352447 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-m6nlk" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.439218 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27"] Dec 17 08:56:35 crc kubenswrapper[4966]: E1217 08:56:35.439718 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f164e75-9cf1-4cde-89ba-32a85cced4b5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.439739 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f164e75-9cf1-4cde-89ba-32a85cced4b5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.440014 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f164e75-9cf1-4cde-89ba-32a85cced4b5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.440651 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.443623 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.444207 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.444498 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.444736 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.461426 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27"] Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.500233 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghkcb\" (UniqueName: \"kubernetes.io/projected/ee245cf4-4e20-44b7-9553-361bac15ce73-kube-api-access-ghkcb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.500317 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.500358 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.602557 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.602742 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghkcb\" (UniqueName: \"kubernetes.io/projected/ee245cf4-4e20-44b7-9553-361bac15ce73-kube-api-access-ghkcb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.602792 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.607594 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.619394 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.623953 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghkcb\" (UniqueName: \"kubernetes.io/projected/ee245cf4-4e20-44b7-9553-361bac15ce73-kube-api-access-ghkcb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:35 crc kubenswrapper[4966]: I1217 08:56:35.757091 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:36 crc kubenswrapper[4966]: I1217 08:56:36.340806 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27"] Dec 17 08:56:36 crc kubenswrapper[4966]: W1217 08:56:36.347532 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee245cf4_4e20_44b7_9553_361bac15ce73.slice/crio-00412b878e442247bac818052d742e57628110be5a2e7666352df0c50affc1c0 WatchSource:0}: Error finding container 00412b878e442247bac818052d742e57628110be5a2e7666352df0c50affc1c0: Status 404 returned error can't find the container with id 00412b878e442247bac818052d742e57628110be5a2e7666352df0c50affc1c0 Dec 17 08:56:36 crc kubenswrapper[4966]: I1217 08:56:36.366226 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" event={"ID":"ee245cf4-4e20-44b7-9553-361bac15ce73","Type":"ContainerStarted","Data":"00412b878e442247bac818052d742e57628110be5a2e7666352df0c50affc1c0"} Dec 17 08:56:39 crc kubenswrapper[4966]: I1217 08:56:39.412146 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" event={"ID":"ee245cf4-4e20-44b7-9553-361bac15ce73","Type":"ContainerStarted","Data":"8eb0e584ee78df18253e482fcbcd55801494a2f5a7220fa4393757f106052861"} Dec 17 08:56:39 crc kubenswrapper[4966]: I1217 08:56:39.433261 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" podStartSLOduration=2.507003304 podStartE2EDuration="4.433244131s" podCreationTimestamp="2025-12-17 08:56:35 +0000 UTC" firstStartedPulling="2025-12-17 08:56:36.35232714 +0000 UTC m=+2131.897397082" lastFinishedPulling="2025-12-17 08:56:38.278567967 +0000 UTC m=+2133.823637909" observedRunningTime="2025-12-17 08:56:39.42736871 +0000 UTC m=+2134.972438642" watchObservedRunningTime="2025-12-17 08:56:39.433244131 +0000 UTC m=+2134.978314073" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.368192 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dpbv6"] Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.374121 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.396262 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dpbv6"] Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.481386 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrtxw\" (UniqueName: \"kubernetes.io/projected/a751940b-c64e-4658-94d9-a84a262911fc-kube-api-access-lrtxw\") pod \"redhat-operators-dpbv6\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.481481 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-catalog-content\") pod \"redhat-operators-dpbv6\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.481511 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-utilities\") pod \"redhat-operators-dpbv6\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.583420 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrtxw\" (UniqueName: \"kubernetes.io/projected/a751940b-c64e-4658-94d9-a84a262911fc-kube-api-access-lrtxw\") pod \"redhat-operators-dpbv6\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.583522 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-catalog-content\") pod \"redhat-operators-dpbv6\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.583556 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-utilities\") pod \"redhat-operators-dpbv6\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.584460 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-utilities\") pod \"redhat-operators-dpbv6\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.584568 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-catalog-content\") pod \"redhat-operators-dpbv6\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.637720 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrtxw\" (UniqueName: \"kubernetes.io/projected/a751940b-c64e-4658-94d9-a84a262911fc-kube-api-access-lrtxw\") pod \"redhat-operators-dpbv6\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:43 crc kubenswrapper[4966]: I1217 08:56:43.703706 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:44 crc kubenswrapper[4966]: I1217 08:56:44.272236 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dpbv6"] Dec 17 08:56:44 crc kubenswrapper[4966]: W1217 08:56:44.284652 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda751940b_c64e_4658_94d9_a84a262911fc.slice/crio-10d9f17aa9800bda65969982127662790738fe44d13379f324e3ffd5d43951fe WatchSource:0}: Error finding container 10d9f17aa9800bda65969982127662790738fe44d13379f324e3ffd5d43951fe: Status 404 returned error can't find the container with id 10d9f17aa9800bda65969982127662790738fe44d13379f324e3ffd5d43951fe Dec 17 08:56:44 crc kubenswrapper[4966]: I1217 08:56:44.462732 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dpbv6" event={"ID":"a751940b-c64e-4658-94d9-a84a262911fc","Type":"ContainerStarted","Data":"10d9f17aa9800bda65969982127662790738fe44d13379f324e3ffd5d43951fe"} Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.164321 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pngfq"] Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.166408 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.189937 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pngfq"] Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.320557 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlvff\" (UniqueName: \"kubernetes.io/projected/9bff8663-2bcf-449a-8a81-e7461bac53db-kube-api-access-wlvff\") pod \"redhat-marketplace-pngfq\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.320808 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-catalog-content\") pod \"redhat-marketplace-pngfq\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.320979 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-utilities\") pod \"redhat-marketplace-pngfq\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.422791 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-catalog-content\") pod \"redhat-marketplace-pngfq\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.422859 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-utilities\") pod \"redhat-marketplace-pngfq\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.423020 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlvff\" (UniqueName: \"kubernetes.io/projected/9bff8663-2bcf-449a-8a81-e7461bac53db-kube-api-access-wlvff\") pod \"redhat-marketplace-pngfq\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.423660 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-utilities\") pod \"redhat-marketplace-pngfq\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.423735 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-catalog-content\") pod \"redhat-marketplace-pngfq\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.443831 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlvff\" (UniqueName: \"kubernetes.io/projected/9bff8663-2bcf-449a-8a81-e7461bac53db-kube-api-access-wlvff\") pod \"redhat-marketplace-pngfq\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.473287 4966 generic.go:334] "Generic (PLEG): container finished" podID="a751940b-c64e-4658-94d9-a84a262911fc" containerID="60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a" exitCode=0 Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.473338 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dpbv6" event={"ID":"a751940b-c64e-4658-94d9-a84a262911fc","Type":"ContainerDied","Data":"60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a"} Dec 17 08:56:45 crc kubenswrapper[4966]: I1217 08:56:45.494368 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:46 crc kubenswrapper[4966]: I1217 08:56:46.043642 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pngfq"] Dec 17 08:56:46 crc kubenswrapper[4966]: W1217 08:56:46.046144 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bff8663_2bcf_449a_8a81_e7461bac53db.slice/crio-1f709ea1dfd7c351310507755e2751fdef2964ec225e267bfbd788c5c5fd0950 WatchSource:0}: Error finding container 1f709ea1dfd7c351310507755e2751fdef2964ec225e267bfbd788c5c5fd0950: Status 404 returned error can't find the container with id 1f709ea1dfd7c351310507755e2751fdef2964ec225e267bfbd788c5c5fd0950 Dec 17 08:56:46 crc kubenswrapper[4966]: I1217 08:56:46.485645 4966 generic.go:334] "Generic (PLEG): container finished" podID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerID="438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889" exitCode=0 Dec 17 08:56:46 crc kubenswrapper[4966]: I1217 08:56:46.485818 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pngfq" event={"ID":"9bff8663-2bcf-449a-8a81-e7461bac53db","Type":"ContainerDied","Data":"438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889"} Dec 17 08:56:46 crc kubenswrapper[4966]: I1217 08:56:46.486064 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pngfq" event={"ID":"9bff8663-2bcf-449a-8a81-e7461bac53db","Type":"ContainerStarted","Data":"1f709ea1dfd7c351310507755e2751fdef2964ec225e267bfbd788c5c5fd0950"} Dec 17 08:56:46 crc kubenswrapper[4966]: I1217 08:56:46.807355 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:56:46 crc kubenswrapper[4966]: I1217 08:56:46.807422 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:56:47 crc kubenswrapper[4966]: I1217 08:56:47.495713 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dpbv6" event={"ID":"a751940b-c64e-4658-94d9-a84a262911fc","Type":"ContainerStarted","Data":"a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207"} Dec 17 08:56:48 crc kubenswrapper[4966]: I1217 08:56:48.506725 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pngfq" event={"ID":"9bff8663-2bcf-449a-8a81-e7461bac53db","Type":"ContainerStarted","Data":"35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90"} Dec 17 08:56:49 crc kubenswrapper[4966]: I1217 08:56:49.519212 4966 generic.go:334] "Generic (PLEG): container finished" podID="ee245cf4-4e20-44b7-9553-361bac15ce73" containerID="8eb0e584ee78df18253e482fcbcd55801494a2f5a7220fa4393757f106052861" exitCode=0 Dec 17 08:56:49 crc kubenswrapper[4966]: I1217 08:56:49.519440 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" event={"ID":"ee245cf4-4e20-44b7-9553-361bac15ce73","Type":"ContainerDied","Data":"8eb0e584ee78df18253e482fcbcd55801494a2f5a7220fa4393757f106052861"} Dec 17 08:56:49 crc kubenswrapper[4966]: I1217 08:56:49.522160 4966 generic.go:334] "Generic (PLEG): container finished" podID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerID="35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90" exitCode=0 Dec 17 08:56:49 crc kubenswrapper[4966]: I1217 08:56:49.522191 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pngfq" event={"ID":"9bff8663-2bcf-449a-8a81-e7461bac53db","Type":"ContainerDied","Data":"35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90"} Dec 17 08:56:50 crc kubenswrapper[4966]: I1217 08:56:50.536669 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pngfq" event={"ID":"9bff8663-2bcf-449a-8a81-e7461bac53db","Type":"ContainerStarted","Data":"602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716"} Dec 17 08:56:50 crc kubenswrapper[4966]: I1217 08:56:50.576778 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pngfq" podStartSLOduration=1.932071275 podStartE2EDuration="5.576755056s" podCreationTimestamp="2025-12-17 08:56:45 +0000 UTC" firstStartedPulling="2025-12-17 08:56:46.487213261 +0000 UTC m=+2142.032283223" lastFinishedPulling="2025-12-17 08:56:50.131897062 +0000 UTC m=+2145.676967004" observedRunningTime="2025-12-17 08:56:50.558479847 +0000 UTC m=+2146.103549799" watchObservedRunningTime="2025-12-17 08:56:50.576755056 +0000 UTC m=+2146.121824998" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.049398 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.247316 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghkcb\" (UniqueName: \"kubernetes.io/projected/ee245cf4-4e20-44b7-9553-361bac15ce73-kube-api-access-ghkcb\") pod \"ee245cf4-4e20-44b7-9553-361bac15ce73\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.247482 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-ssh-key\") pod \"ee245cf4-4e20-44b7-9553-361bac15ce73\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.247566 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-inventory\") pod \"ee245cf4-4e20-44b7-9553-361bac15ce73\" (UID: \"ee245cf4-4e20-44b7-9553-361bac15ce73\") " Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.285108 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee245cf4-4e20-44b7-9553-361bac15ce73-kube-api-access-ghkcb" (OuterVolumeSpecName: "kube-api-access-ghkcb") pod "ee245cf4-4e20-44b7-9553-361bac15ce73" (UID: "ee245cf4-4e20-44b7-9553-361bac15ce73"). InnerVolumeSpecName "kube-api-access-ghkcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.305027 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-inventory" (OuterVolumeSpecName: "inventory") pod "ee245cf4-4e20-44b7-9553-361bac15ce73" (UID: "ee245cf4-4e20-44b7-9553-361bac15ce73"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.314818 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ee245cf4-4e20-44b7-9553-361bac15ce73" (UID: "ee245cf4-4e20-44b7-9553-361bac15ce73"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.349644 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghkcb\" (UniqueName: \"kubernetes.io/projected/ee245cf4-4e20-44b7-9553-361bac15ce73-kube-api-access-ghkcb\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.349674 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.349683 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee245cf4-4e20-44b7-9553-361bac15ce73-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.547239 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" event={"ID":"ee245cf4-4e20-44b7-9553-361bac15ce73","Type":"ContainerDied","Data":"00412b878e442247bac818052d742e57628110be5a2e7666352df0c50affc1c0"} Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.547285 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00412b878e442247bac818052d742e57628110be5a2e7666352df0c50affc1c0" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.547252 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5s27" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.549795 4966 generic.go:334] "Generic (PLEG): container finished" podID="a751940b-c64e-4658-94d9-a84a262911fc" containerID="a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207" exitCode=0 Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.550406 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dpbv6" event={"ID":"a751940b-c64e-4658-94d9-a84a262911fc","Type":"ContainerDied","Data":"a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207"} Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.663566 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn"] Dec 17 08:56:51 crc kubenswrapper[4966]: E1217 08:56:51.664552 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee245cf4-4e20-44b7-9553-361bac15ce73" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.664645 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee245cf4-4e20-44b7-9553-361bac15ce73" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.664906 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee245cf4-4e20-44b7-9553-361bac15ce73" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.665563 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.669538 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.669751 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.670436 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.670581 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.670750 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.670998 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.671212 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.674330 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.689567 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn"] Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.865681 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.865771 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.865808 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.865851 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.865893 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.865924 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.865948 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.865968 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7l86\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-kube-api-access-r7l86\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.866003 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.866744 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.866927 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.866998 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.867028 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.867062 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969530 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969577 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969602 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969636 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969686 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969714 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969767 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969792 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969819 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969839 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969861 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7l86\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-kube-api-access-r7l86\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969906 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969946 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.969983 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.974100 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.974111 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.974285 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.977159 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.978270 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.978832 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.978959 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.980251 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.981473 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.983296 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.983791 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.984865 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.988352 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:51 crc kubenswrapper[4966]: I1217 08:56:51.990977 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7l86\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-kube-api-access-r7l86\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:52 crc kubenswrapper[4966]: I1217 08:56:52.284739 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:56:52 crc kubenswrapper[4966]: I1217 08:56:52.879943 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn"] Dec 17 08:56:53 crc kubenswrapper[4966]: I1217 08:56:53.566711 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" event={"ID":"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7","Type":"ContainerStarted","Data":"1d3cadc0b239ef164c21844ac57dc04aea2f6264a27b0c6c24399f587ad7d807"} Dec 17 08:56:53 crc kubenswrapper[4966]: I1217 08:56:53.568817 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dpbv6" event={"ID":"a751940b-c64e-4658-94d9-a84a262911fc","Type":"ContainerStarted","Data":"b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514"} Dec 17 08:56:53 crc kubenswrapper[4966]: I1217 08:56:53.705421 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:53 crc kubenswrapper[4966]: I1217 08:56:53.705830 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:56:54 crc kubenswrapper[4966]: I1217 08:56:54.580542 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" event={"ID":"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7","Type":"ContainerStarted","Data":"9d52247eaccb29a47cbacd6fe8c52d1118416d2a0e16e7d80bc7c4b51d3b9c25"} Dec 17 08:56:54 crc kubenswrapper[4966]: I1217 08:56:54.607254 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" podStartSLOduration=2.754966753 podStartE2EDuration="3.607232719s" podCreationTimestamp="2025-12-17 08:56:51 +0000 UTC" firstStartedPulling="2025-12-17 08:56:52.888570231 +0000 UTC m=+2148.433640173" lastFinishedPulling="2025-12-17 08:56:53.740836197 +0000 UTC m=+2149.285906139" observedRunningTime="2025-12-17 08:56:54.60107498 +0000 UTC m=+2150.146144932" watchObservedRunningTime="2025-12-17 08:56:54.607232719 +0000 UTC m=+2150.152302661" Dec 17 08:56:54 crc kubenswrapper[4966]: I1217 08:56:54.611122 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dpbv6" podStartSLOduration=4.6611938219999995 podStartE2EDuration="11.611107724s" podCreationTimestamp="2025-12-17 08:56:43 +0000 UTC" firstStartedPulling="2025-12-17 08:56:45.476161745 +0000 UTC m=+2141.021231687" lastFinishedPulling="2025-12-17 08:56:52.426075647 +0000 UTC m=+2147.971145589" observedRunningTime="2025-12-17 08:56:53.587909106 +0000 UTC m=+2149.132979078" watchObservedRunningTime="2025-12-17 08:56:54.611107724 +0000 UTC m=+2150.156177656" Dec 17 08:56:54 crc kubenswrapper[4966]: I1217 08:56:54.774698 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dpbv6" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="registry-server" probeResult="failure" output=< Dec 17 08:56:54 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:56:54 crc kubenswrapper[4966]: > Dec 17 08:56:55 crc kubenswrapper[4966]: I1217 08:56:55.495582 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:55 crc kubenswrapper[4966]: I1217 08:56:55.495636 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:55 crc kubenswrapper[4966]: I1217 08:56:55.556927 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:55 crc kubenswrapper[4966]: I1217 08:56:55.663860 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:55 crc kubenswrapper[4966]: I1217 08:56:55.955408 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pngfq"] Dec 17 08:56:57 crc kubenswrapper[4966]: I1217 08:56:57.611163 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pngfq" podUID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerName="registry-server" containerID="cri-o://602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716" gracePeriod=2 Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.128189 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.196563 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-utilities\") pod \"9bff8663-2bcf-449a-8a81-e7461bac53db\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.197109 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-catalog-content\") pod \"9bff8663-2bcf-449a-8a81-e7461bac53db\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.197266 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlvff\" (UniqueName: \"kubernetes.io/projected/9bff8663-2bcf-449a-8a81-e7461bac53db-kube-api-access-wlvff\") pod \"9bff8663-2bcf-449a-8a81-e7461bac53db\" (UID: \"9bff8663-2bcf-449a-8a81-e7461bac53db\") " Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.197401 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-utilities" (OuterVolumeSpecName: "utilities") pod "9bff8663-2bcf-449a-8a81-e7461bac53db" (UID: "9bff8663-2bcf-449a-8a81-e7461bac53db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.199042 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.202763 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bff8663-2bcf-449a-8a81-e7461bac53db-kube-api-access-wlvff" (OuterVolumeSpecName: "kube-api-access-wlvff") pod "9bff8663-2bcf-449a-8a81-e7461bac53db" (UID: "9bff8663-2bcf-449a-8a81-e7461bac53db"). InnerVolumeSpecName "kube-api-access-wlvff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.219865 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bff8663-2bcf-449a-8a81-e7461bac53db" (UID: "9bff8663-2bcf-449a-8a81-e7461bac53db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.301061 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bff8663-2bcf-449a-8a81-e7461bac53db-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.301108 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlvff\" (UniqueName: \"kubernetes.io/projected/9bff8663-2bcf-449a-8a81-e7461bac53db-kube-api-access-wlvff\") on node \"crc\" DevicePath \"\"" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.636749 4966 generic.go:334] "Generic (PLEG): container finished" podID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerID="602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716" exitCode=0 Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.636845 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pngfq" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.636849 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pngfq" event={"ID":"9bff8663-2bcf-449a-8a81-e7461bac53db","Type":"ContainerDied","Data":"602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716"} Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.638183 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pngfq" event={"ID":"9bff8663-2bcf-449a-8a81-e7461bac53db","Type":"ContainerDied","Data":"1f709ea1dfd7c351310507755e2751fdef2964ec225e267bfbd788c5c5fd0950"} Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.638215 4966 scope.go:117] "RemoveContainer" containerID="602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.678261 4966 scope.go:117] "RemoveContainer" containerID="35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.687284 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pngfq"] Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.696476 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pngfq"] Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.733110 4966 scope.go:117] "RemoveContainer" containerID="438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.781796 4966 scope.go:117] "RemoveContainer" containerID="602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716" Dec 17 08:56:58 crc kubenswrapper[4966]: E1217 08:56:58.782269 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716\": container with ID starting with 602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716 not found: ID does not exist" containerID="602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.782325 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716"} err="failed to get container status \"602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716\": rpc error: code = NotFound desc = could not find container \"602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716\": container with ID starting with 602d026d88428d1abb9a4a8b596933feca4aca4d87334945fba406f043c56716 not found: ID does not exist" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.782353 4966 scope.go:117] "RemoveContainer" containerID="35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90" Dec 17 08:56:58 crc kubenswrapper[4966]: E1217 08:56:58.782711 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90\": container with ID starting with 35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90 not found: ID does not exist" containerID="35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.782742 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90"} err="failed to get container status \"35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90\": rpc error: code = NotFound desc = could not find container \"35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90\": container with ID starting with 35baba3f93a60a6ff68bbc2ca61cb7237fac7a414501e9a77d77555507cbef90 not found: ID does not exist" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.782763 4966 scope.go:117] "RemoveContainer" containerID="438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889" Dec 17 08:56:58 crc kubenswrapper[4966]: E1217 08:56:58.783169 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889\": container with ID starting with 438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889 not found: ID does not exist" containerID="438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.783193 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889"} err="failed to get container status \"438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889\": rpc error: code = NotFound desc = could not find container \"438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889\": container with ID starting with 438ce785063f17d2f4675ed6219508f1768f267661ebf1ba4af0fad8a6e52889 not found: ID does not exist" Dec 17 08:56:58 crc kubenswrapper[4966]: I1217 08:56:58.843045 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bff8663-2bcf-449a-8a81-e7461bac53db" path="/var/lib/kubelet/pods/9bff8663-2bcf-449a-8a81-e7461bac53db/volumes" Dec 17 08:57:04 crc kubenswrapper[4966]: I1217 08:57:04.758684 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dpbv6" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="registry-server" probeResult="failure" output=< Dec 17 08:57:04 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 08:57:04 crc kubenswrapper[4966]: > Dec 17 08:57:13 crc kubenswrapper[4966]: I1217 08:57:13.760907 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:57:13 crc kubenswrapper[4966]: I1217 08:57:13.814784 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:57:14 crc kubenswrapper[4966]: I1217 08:57:14.569396 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dpbv6"] Dec 17 08:57:14 crc kubenswrapper[4966]: I1217 08:57:14.809089 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dpbv6" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="registry-server" containerID="cri-o://b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514" gracePeriod=2 Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.323750 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.441807 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-utilities\") pod \"a751940b-c64e-4658-94d9-a84a262911fc\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.442087 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrtxw\" (UniqueName: \"kubernetes.io/projected/a751940b-c64e-4658-94d9-a84a262911fc-kube-api-access-lrtxw\") pod \"a751940b-c64e-4658-94d9-a84a262911fc\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.442270 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-catalog-content\") pod \"a751940b-c64e-4658-94d9-a84a262911fc\" (UID: \"a751940b-c64e-4658-94d9-a84a262911fc\") " Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.442939 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-utilities" (OuterVolumeSpecName: "utilities") pod "a751940b-c64e-4658-94d9-a84a262911fc" (UID: "a751940b-c64e-4658-94d9-a84a262911fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.443390 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.452275 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a751940b-c64e-4658-94d9-a84a262911fc-kube-api-access-lrtxw" (OuterVolumeSpecName: "kube-api-access-lrtxw") pod "a751940b-c64e-4658-94d9-a84a262911fc" (UID: "a751940b-c64e-4658-94d9-a84a262911fc"). InnerVolumeSpecName "kube-api-access-lrtxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.545553 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrtxw\" (UniqueName: \"kubernetes.io/projected/a751940b-c64e-4658-94d9-a84a262911fc-kube-api-access-lrtxw\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.557559 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a751940b-c64e-4658-94d9-a84a262911fc" (UID: "a751940b-c64e-4658-94d9-a84a262911fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.647504 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a751940b-c64e-4658-94d9-a84a262911fc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.822001 4966 generic.go:334] "Generic (PLEG): container finished" podID="a751940b-c64e-4658-94d9-a84a262911fc" containerID="b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514" exitCode=0 Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.822087 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dpbv6" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.822114 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dpbv6" event={"ID":"a751940b-c64e-4658-94d9-a84a262911fc","Type":"ContainerDied","Data":"b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514"} Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.823176 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dpbv6" event={"ID":"a751940b-c64e-4658-94d9-a84a262911fc","Type":"ContainerDied","Data":"10d9f17aa9800bda65969982127662790738fe44d13379f324e3ffd5d43951fe"} Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.823204 4966 scope.go:117] "RemoveContainer" containerID="b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.860257 4966 scope.go:117] "RemoveContainer" containerID="a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.895584 4966 scope.go:117] "RemoveContainer" containerID="60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.907062 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dpbv6"] Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.921235 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dpbv6"] Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.961137 4966 scope.go:117] "RemoveContainer" containerID="b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514" Dec 17 08:57:15 crc kubenswrapper[4966]: E1217 08:57:15.961603 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514\": container with ID starting with b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514 not found: ID does not exist" containerID="b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.961633 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514"} err="failed to get container status \"b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514\": rpc error: code = NotFound desc = could not find container \"b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514\": container with ID starting with b948cb5c8dc41a535ca4e65435d2abb3a9dc90bb227a73e47e6e122df2441514 not found: ID does not exist" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.961656 4966 scope.go:117] "RemoveContainer" containerID="a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207" Dec 17 08:57:15 crc kubenswrapper[4966]: E1217 08:57:15.961901 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207\": container with ID starting with a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207 not found: ID does not exist" containerID="a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.961929 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207"} err="failed to get container status \"a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207\": rpc error: code = NotFound desc = could not find container \"a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207\": container with ID starting with a35b37c5f3467f5cbe3de51c530f9815307759c6ee0763a2745796f4b2841207 not found: ID does not exist" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.961945 4966 scope.go:117] "RemoveContainer" containerID="60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a" Dec 17 08:57:15 crc kubenswrapper[4966]: E1217 08:57:15.962182 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a\": container with ID starting with 60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a not found: ID does not exist" containerID="60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a" Dec 17 08:57:15 crc kubenswrapper[4966]: I1217 08:57:15.962203 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a"} err="failed to get container status \"60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a\": rpc error: code = NotFound desc = could not find container \"60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a\": container with ID starting with 60b589c015a836dcf6f0490a5ae1a88d8f2ec0fe294fbe16b1b2c5cb2cf3f03a not found: ID does not exist" Dec 17 08:57:16 crc kubenswrapper[4966]: I1217 08:57:16.807254 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:57:16 crc kubenswrapper[4966]: I1217 08:57:16.807608 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:57:16 crc kubenswrapper[4966]: I1217 08:57:16.842276 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a751940b-c64e-4658-94d9-a84a262911fc" path="/var/lib/kubelet/pods/a751940b-c64e-4658-94d9-a84a262911fc/volumes" Dec 17 08:57:35 crc kubenswrapper[4966]: I1217 08:57:35.007778 4966 generic.go:334] "Generic (PLEG): container finished" podID="fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" containerID="9d52247eaccb29a47cbacd6fe8c52d1118416d2a0e16e7d80bc7c4b51d3b9c25" exitCode=0 Dec 17 08:57:35 crc kubenswrapper[4966]: I1217 08:57:35.007847 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" event={"ID":"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7","Type":"ContainerDied","Data":"9d52247eaccb29a47cbacd6fe8c52d1118416d2a0e16e7d80bc7c4b51d3b9c25"} Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.501074 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595074 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595134 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-repo-setup-combined-ca-bundle\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595221 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-inventory\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595253 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595327 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7l86\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-kube-api-access-r7l86\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595342 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-bootstrap-combined-ca-bundle\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595408 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-telemetry-combined-ca-bundle\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595439 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-nova-combined-ca-bundle\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595458 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ovn-combined-ca-bundle\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595474 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-neutron-metadata-combined-ca-bundle\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595494 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595524 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595609 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ssh-key\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.595626 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-libvirt-combined-ca-bundle\") pod \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\" (UID: \"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7\") " Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.603491 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.604248 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.604377 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-kube-api-access-r7l86" (OuterVolumeSpecName: "kube-api-access-r7l86") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "kube-api-access-r7l86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.604778 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.604828 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.605500 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.605933 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.607284 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.607363 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.607729 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.608797 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.614248 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.629988 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-inventory" (OuterVolumeSpecName: "inventory") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.632655 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" (UID: "fb36c36b-a56f-425f-a7a8-1fbb1235b0c7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697330 4966 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697383 4966 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697397 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697411 4966 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697424 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7l86\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-kube-api-access-r7l86\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697433 4966 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697442 4966 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697450 4966 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697458 4966 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697466 4966 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697475 4966 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697485 4966 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697493 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:36 crc kubenswrapper[4966]: I1217 08:57:36.697503 4966 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb36c36b-a56f-425f-a7a8-1fbb1235b0c7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.025447 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" event={"ID":"fb36c36b-a56f-425f-a7a8-1fbb1235b0c7","Type":"ContainerDied","Data":"1d3cadc0b239ef164c21844ac57dc04aea2f6264a27b0c6c24399f587ad7d807"} Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.025682 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d3cadc0b239ef164c21844ac57dc04aea2f6264a27b0c6c24399f587ad7d807" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.025478 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4qlpn" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.136613 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb"] Dec 17 08:57:37 crc kubenswrapper[4966]: E1217 08:57:37.137066 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerName="registry-server" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137082 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerName="registry-server" Dec 17 08:57:37 crc kubenswrapper[4966]: E1217 08:57:37.137103 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="extract-content" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137110 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="extract-content" Dec 17 08:57:37 crc kubenswrapper[4966]: E1217 08:57:37.137129 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137136 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 17 08:57:37 crc kubenswrapper[4966]: E1217 08:57:37.137148 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerName="extract-utilities" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137154 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerName="extract-utilities" Dec 17 08:57:37 crc kubenswrapper[4966]: E1217 08:57:37.137161 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerName="extract-content" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137167 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerName="extract-content" Dec 17 08:57:37 crc kubenswrapper[4966]: E1217 08:57:37.137181 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="extract-utilities" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137186 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="extract-utilities" Dec 17 08:57:37 crc kubenswrapper[4966]: E1217 08:57:37.137197 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="registry-server" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137202 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="registry-server" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137384 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bff8663-2bcf-449a-8a81-e7461bac53db" containerName="registry-server" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137397 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb36c36b-a56f-425f-a7a8-1fbb1235b0c7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.137411 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a751940b-c64e-4658-94d9-a84a262911fc" containerName="registry-server" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.138129 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.140609 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.140652 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.141378 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.141596 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.146587 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb"] Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.146993 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.206522 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.206769 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.207019 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c813431d-6629-4804-bf18-52c0f9634050-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.207101 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.207202 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz4zt\" (UniqueName: \"kubernetes.io/projected/c813431d-6629-4804-bf18-52c0f9634050-kube-api-access-rz4zt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.310204 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.310463 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.310631 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c813431d-6629-4804-bf18-52c0f9634050-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.310757 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.310923 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz4zt\" (UniqueName: \"kubernetes.io/projected/c813431d-6629-4804-bf18-52c0f9634050-kube-api-access-rz4zt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.312416 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c813431d-6629-4804-bf18-52c0f9634050-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.321686 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.322491 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.327765 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.328214 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz4zt\" (UniqueName: \"kubernetes.io/projected/c813431d-6629-4804-bf18-52c0f9634050-kube-api-access-rz4zt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-g5kkb\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:37 crc kubenswrapper[4966]: I1217 08:57:37.456237 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:57:38 crc kubenswrapper[4966]: I1217 08:57:38.048378 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb"] Dec 17 08:57:38 crc kubenswrapper[4966]: I1217 08:57:38.056166 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 08:57:39 crc kubenswrapper[4966]: I1217 08:57:39.046480 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" event={"ID":"c813431d-6629-4804-bf18-52c0f9634050","Type":"ContainerStarted","Data":"2a1c8952ef045db4b00edd473dbd0ec3afaded548b0655aa97458d364ef7ee81"} Dec 17 08:57:40 crc kubenswrapper[4966]: I1217 08:57:40.059792 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" event={"ID":"c813431d-6629-4804-bf18-52c0f9634050","Type":"ContainerStarted","Data":"1b4aa098c6d1e2509350decdca7130996eed91d5ea387c7d035cc3ff67d85d43"} Dec 17 08:57:40 crc kubenswrapper[4966]: I1217 08:57:40.085381 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" podStartSLOduration=2.301021328 podStartE2EDuration="3.085363191s" podCreationTimestamp="2025-12-17 08:57:37 +0000 UTC" firstStartedPulling="2025-12-17 08:57:38.055944317 +0000 UTC m=+2193.601014259" lastFinishedPulling="2025-12-17 08:57:38.84028618 +0000 UTC m=+2194.385356122" observedRunningTime="2025-12-17 08:57:40.076337324 +0000 UTC m=+2195.621407276" watchObservedRunningTime="2025-12-17 08:57:40.085363191 +0000 UTC m=+2195.630433133" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.766486 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tt6w4"] Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.768889 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.798794 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt6w4"] Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.814326 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.814368 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.814418 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.815092 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4757810d46cf706e3c032dafdc0e43a6e21fb930b0834212015c013294922d1f"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.815147 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://4757810d46cf706e3c032dafdc0e43a6e21fb930b0834212015c013294922d1f" gracePeriod=600 Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.829309 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-utilities\") pod \"certified-operators-tt6w4\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.829465 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g95r4\" (UniqueName: \"kubernetes.io/projected/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-kube-api-access-g95r4\") pod \"certified-operators-tt6w4\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.829561 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-catalog-content\") pod \"certified-operators-tt6w4\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.932729 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-utilities\") pod \"certified-operators-tt6w4\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.932985 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g95r4\" (UniqueName: \"kubernetes.io/projected/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-kube-api-access-g95r4\") pod \"certified-operators-tt6w4\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.933466 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-catalog-content\") pod \"certified-operators-tt6w4\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.940285 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-utilities\") pod \"certified-operators-tt6w4\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.941709 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-catalog-content\") pod \"certified-operators-tt6w4\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:46 crc kubenswrapper[4966]: I1217 08:57:46.955092 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g95r4\" (UniqueName: \"kubernetes.io/projected/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-kube-api-access-g95r4\") pod \"certified-operators-tt6w4\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:47 crc kubenswrapper[4966]: I1217 08:57:47.117577 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:47 crc kubenswrapper[4966]: I1217 08:57:47.130686 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="4757810d46cf706e3c032dafdc0e43a6e21fb930b0834212015c013294922d1f" exitCode=0 Dec 17 08:57:47 crc kubenswrapper[4966]: I1217 08:57:47.130729 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"4757810d46cf706e3c032dafdc0e43a6e21fb930b0834212015c013294922d1f"} Dec 17 08:57:47 crc kubenswrapper[4966]: I1217 08:57:47.130767 4966 scope.go:117] "RemoveContainer" containerID="927d999d75ab21766267795cf18c6b8dd9166cfcdbc7f5250359629d5523c48b" Dec 17 08:57:47 crc kubenswrapper[4966]: I1217 08:57:47.643436 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tt6w4"] Dec 17 08:57:48 crc kubenswrapper[4966]: I1217 08:57:48.140694 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592"} Dec 17 08:57:48 crc kubenswrapper[4966]: I1217 08:57:48.142731 4966 generic.go:334] "Generic (PLEG): container finished" podID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerID="1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2" exitCode=0 Dec 17 08:57:48 crc kubenswrapper[4966]: I1217 08:57:48.142773 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6w4" event={"ID":"5669feb1-1d9f-4629-a1ed-f1f4edf539e2","Type":"ContainerDied","Data":"1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2"} Dec 17 08:57:48 crc kubenswrapper[4966]: I1217 08:57:48.142797 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6w4" event={"ID":"5669feb1-1d9f-4629-a1ed-f1f4edf539e2","Type":"ContainerStarted","Data":"1c989cca52abfa81cffa3ba748b4f279879b00c1b6b97720d82afa4d76c3867b"} Dec 17 08:57:50 crc kubenswrapper[4966]: I1217 08:57:50.162328 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6w4" event={"ID":"5669feb1-1d9f-4629-a1ed-f1f4edf539e2","Type":"ContainerStarted","Data":"10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130"} Dec 17 08:57:51 crc kubenswrapper[4966]: I1217 08:57:51.171658 4966 generic.go:334] "Generic (PLEG): container finished" podID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerID="10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130" exitCode=0 Dec 17 08:57:51 crc kubenswrapper[4966]: I1217 08:57:51.171703 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6w4" event={"ID":"5669feb1-1d9f-4629-a1ed-f1f4edf539e2","Type":"ContainerDied","Data":"10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130"} Dec 17 08:57:52 crc kubenswrapper[4966]: I1217 08:57:52.182095 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6w4" event={"ID":"5669feb1-1d9f-4629-a1ed-f1f4edf539e2","Type":"ContainerStarted","Data":"91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635"} Dec 17 08:57:52 crc kubenswrapper[4966]: I1217 08:57:52.205060 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tt6w4" podStartSLOduration=3.603616354 podStartE2EDuration="6.205042509s" podCreationTimestamp="2025-12-17 08:57:46 +0000 UTC" firstStartedPulling="2025-12-17 08:57:49.152421477 +0000 UTC m=+2204.697491419" lastFinishedPulling="2025-12-17 08:57:51.753847632 +0000 UTC m=+2207.298917574" observedRunningTime="2025-12-17 08:57:52.197255876 +0000 UTC m=+2207.742325818" watchObservedRunningTime="2025-12-17 08:57:52.205042509 +0000 UTC m=+2207.750112451" Dec 17 08:57:57 crc kubenswrapper[4966]: I1217 08:57:57.118036 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:57 crc kubenswrapper[4966]: I1217 08:57:57.118463 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:57 crc kubenswrapper[4966]: I1217 08:57:57.163118 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:57 crc kubenswrapper[4966]: I1217 08:57:57.274993 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:57 crc kubenswrapper[4966]: I1217 08:57:57.416149 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tt6w4"] Dec 17 08:57:59 crc kubenswrapper[4966]: I1217 08:57:59.252421 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tt6w4" podUID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerName="registry-server" containerID="cri-o://91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635" gracePeriod=2 Dec 17 08:57:59 crc kubenswrapper[4966]: I1217 08:57:59.760292 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:57:59 crc kubenswrapper[4966]: I1217 08:57:59.907723 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-utilities\") pod \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " Dec 17 08:57:59 crc kubenswrapper[4966]: I1217 08:57:59.907957 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-catalog-content\") pod \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " Dec 17 08:57:59 crc kubenswrapper[4966]: I1217 08:57:59.907982 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g95r4\" (UniqueName: \"kubernetes.io/projected/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-kube-api-access-g95r4\") pod \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\" (UID: \"5669feb1-1d9f-4629-a1ed-f1f4edf539e2\") " Dec 17 08:57:59 crc kubenswrapper[4966]: I1217 08:57:59.910222 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-utilities" (OuterVolumeSpecName: "utilities") pod "5669feb1-1d9f-4629-a1ed-f1f4edf539e2" (UID: "5669feb1-1d9f-4629-a1ed-f1f4edf539e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:57:59 crc kubenswrapper[4966]: I1217 08:57:59.915637 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-kube-api-access-g95r4" (OuterVolumeSpecName: "kube-api-access-g95r4") pod "5669feb1-1d9f-4629-a1ed-f1f4edf539e2" (UID: "5669feb1-1d9f-4629-a1ed-f1f4edf539e2"). InnerVolumeSpecName "kube-api-access-g95r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:57:59 crc kubenswrapper[4966]: I1217 08:57:59.976968 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5669feb1-1d9f-4629-a1ed-f1f4edf539e2" (UID: "5669feb1-1d9f-4629-a1ed-f1f4edf539e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.010103 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.010134 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.010146 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g95r4\" (UniqueName: \"kubernetes.io/projected/5669feb1-1d9f-4629-a1ed-f1f4edf539e2-kube-api-access-g95r4\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.266548 4966 generic.go:334] "Generic (PLEG): container finished" podID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerID="91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635" exitCode=0 Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.266606 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6w4" event={"ID":"5669feb1-1d9f-4629-a1ed-f1f4edf539e2","Type":"ContainerDied","Data":"91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635"} Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.266654 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tt6w4" event={"ID":"5669feb1-1d9f-4629-a1ed-f1f4edf539e2","Type":"ContainerDied","Data":"1c989cca52abfa81cffa3ba748b4f279879b00c1b6b97720d82afa4d76c3867b"} Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.266677 4966 scope.go:117] "RemoveContainer" containerID="91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.266758 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tt6w4" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.298832 4966 scope.go:117] "RemoveContainer" containerID="10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.343820 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tt6w4"] Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.351575 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tt6w4"] Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.356108 4966 scope.go:117] "RemoveContainer" containerID="1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.392699 4966 scope.go:117] "RemoveContainer" containerID="91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635" Dec 17 08:58:00 crc kubenswrapper[4966]: E1217 08:58:00.393224 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635\": container with ID starting with 91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635 not found: ID does not exist" containerID="91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.393269 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635"} err="failed to get container status \"91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635\": rpc error: code = NotFound desc = could not find container \"91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635\": container with ID starting with 91f8afb0ccf8ff026bf98d06b0aa04a9763f4b1d702ec70903043ee66cc3d635 not found: ID does not exist" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.393296 4966 scope.go:117] "RemoveContainer" containerID="10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130" Dec 17 08:58:00 crc kubenswrapper[4966]: E1217 08:58:00.393529 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130\": container with ID starting with 10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130 not found: ID does not exist" containerID="10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.393558 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130"} err="failed to get container status \"10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130\": rpc error: code = NotFound desc = could not find container \"10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130\": container with ID starting with 10d8f08022bb831be3d7120683053f113ad66fb81f621e4bbbe62ee485332130 not found: ID does not exist" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.393572 4966 scope.go:117] "RemoveContainer" containerID="1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2" Dec 17 08:58:00 crc kubenswrapper[4966]: E1217 08:58:00.393906 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2\": container with ID starting with 1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2 not found: ID does not exist" containerID="1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.393936 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2"} err="failed to get container status \"1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2\": rpc error: code = NotFound desc = could not find container \"1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2\": container with ID starting with 1f0e42333431baab45827995a37d2d6afd8111bb6eff0a369d4cfc239a94e2e2 not found: ID does not exist" Dec 17 08:58:00 crc kubenswrapper[4966]: I1217 08:58:00.843980 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" path="/var/lib/kubelet/pods/5669feb1-1d9f-4629-a1ed-f1f4edf539e2/volumes" Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.816695 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x5qdp"] Dec 17 08:58:09 crc kubenswrapper[4966]: E1217 08:58:09.817847 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerName="registry-server" Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.817863 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerName="registry-server" Dec 17 08:58:09 crc kubenswrapper[4966]: E1217 08:58:09.817907 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerName="extract-content" Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.817916 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerName="extract-content" Dec 17 08:58:09 crc kubenswrapper[4966]: E1217 08:58:09.817947 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerName="extract-utilities" Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.817956 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerName="extract-utilities" Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.818206 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="5669feb1-1d9f-4629-a1ed-f1f4edf539e2" containerName="registry-server" Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.819936 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.841185 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x5qdp"] Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.917540 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-utilities\") pod \"community-operators-x5qdp\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.917637 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-catalog-content\") pod \"community-operators-x5qdp\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:09 crc kubenswrapper[4966]: I1217 08:58:09.917951 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bghgk\" (UniqueName: \"kubernetes.io/projected/bea1b22d-72c0-4558-845c-52ea15d2eb94-kube-api-access-bghgk\") pod \"community-operators-x5qdp\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:10 crc kubenswrapper[4966]: I1217 08:58:10.019463 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-utilities\") pod \"community-operators-x5qdp\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:10 crc kubenswrapper[4966]: I1217 08:58:10.019522 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-catalog-content\") pod \"community-operators-x5qdp\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:10 crc kubenswrapper[4966]: I1217 08:58:10.019583 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bghgk\" (UniqueName: \"kubernetes.io/projected/bea1b22d-72c0-4558-845c-52ea15d2eb94-kube-api-access-bghgk\") pod \"community-operators-x5qdp\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:10 crc kubenswrapper[4966]: I1217 08:58:10.020186 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-catalog-content\") pod \"community-operators-x5qdp\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:10 crc kubenswrapper[4966]: I1217 08:58:10.020196 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-utilities\") pod \"community-operators-x5qdp\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:10 crc kubenswrapper[4966]: I1217 08:58:10.053848 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bghgk\" (UniqueName: \"kubernetes.io/projected/bea1b22d-72c0-4558-845c-52ea15d2eb94-kube-api-access-bghgk\") pod \"community-operators-x5qdp\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:10 crc kubenswrapper[4966]: I1217 08:58:10.139355 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:10 crc kubenswrapper[4966]: I1217 08:58:10.692510 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x5qdp"] Dec 17 08:58:10 crc kubenswrapper[4966]: W1217 08:58:10.694057 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbea1b22d_72c0_4558_845c_52ea15d2eb94.slice/crio-9e07d9c3db92f403b5539e910419b1c33f9820488b8e558d4ebccabdeaee76d4 WatchSource:0}: Error finding container 9e07d9c3db92f403b5539e910419b1c33f9820488b8e558d4ebccabdeaee76d4: Status 404 returned error can't find the container with id 9e07d9c3db92f403b5539e910419b1c33f9820488b8e558d4ebccabdeaee76d4 Dec 17 08:58:11 crc kubenswrapper[4966]: I1217 08:58:11.388203 4966 generic.go:334] "Generic (PLEG): container finished" podID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerID="10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a" exitCode=0 Dec 17 08:58:11 crc kubenswrapper[4966]: I1217 08:58:11.388418 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5qdp" event={"ID":"bea1b22d-72c0-4558-845c-52ea15d2eb94","Type":"ContainerDied","Data":"10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a"} Dec 17 08:58:11 crc kubenswrapper[4966]: I1217 08:58:11.392637 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5qdp" event={"ID":"bea1b22d-72c0-4558-845c-52ea15d2eb94","Type":"ContainerStarted","Data":"9e07d9c3db92f403b5539e910419b1c33f9820488b8e558d4ebccabdeaee76d4"} Dec 17 08:58:12 crc kubenswrapper[4966]: I1217 08:58:12.403255 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5qdp" event={"ID":"bea1b22d-72c0-4558-845c-52ea15d2eb94","Type":"ContainerStarted","Data":"392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8"} Dec 17 08:58:13 crc kubenswrapper[4966]: I1217 08:58:13.677970 4966 generic.go:334] "Generic (PLEG): container finished" podID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerID="392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8" exitCode=0 Dec 17 08:58:13 crc kubenswrapper[4966]: I1217 08:58:13.678043 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5qdp" event={"ID":"bea1b22d-72c0-4558-845c-52ea15d2eb94","Type":"ContainerDied","Data":"392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8"} Dec 17 08:58:14 crc kubenswrapper[4966]: I1217 08:58:14.689745 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5qdp" event={"ID":"bea1b22d-72c0-4558-845c-52ea15d2eb94","Type":"ContainerStarted","Data":"0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa"} Dec 17 08:58:14 crc kubenswrapper[4966]: I1217 08:58:14.710861 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x5qdp" podStartSLOduration=2.853488059 podStartE2EDuration="5.710840613s" podCreationTimestamp="2025-12-17 08:58:09 +0000 UTC" firstStartedPulling="2025-12-17 08:58:11.391072455 +0000 UTC m=+2226.936142397" lastFinishedPulling="2025-12-17 08:58:14.248424989 +0000 UTC m=+2229.793494951" observedRunningTime="2025-12-17 08:58:14.709982749 +0000 UTC m=+2230.255052701" watchObservedRunningTime="2025-12-17 08:58:14.710840613 +0000 UTC m=+2230.255910565" Dec 17 08:58:20 crc kubenswrapper[4966]: I1217 08:58:20.139824 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:20 crc kubenswrapper[4966]: I1217 08:58:20.140524 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:20 crc kubenswrapper[4966]: I1217 08:58:20.190178 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:20 crc kubenswrapper[4966]: I1217 08:58:20.804498 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:20 crc kubenswrapper[4966]: I1217 08:58:20.903109 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x5qdp"] Dec 17 08:58:22 crc kubenswrapper[4966]: I1217 08:58:22.761775 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x5qdp" podUID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerName="registry-server" containerID="cri-o://0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa" gracePeriod=2 Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.718404 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.777748 4966 generic.go:334] "Generic (PLEG): container finished" podID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerID="0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa" exitCode=0 Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.777787 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5qdp" event={"ID":"bea1b22d-72c0-4558-845c-52ea15d2eb94","Type":"ContainerDied","Data":"0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa"} Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.777810 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5qdp" event={"ID":"bea1b22d-72c0-4558-845c-52ea15d2eb94","Type":"ContainerDied","Data":"9e07d9c3db92f403b5539e910419b1c33f9820488b8e558d4ebccabdeaee76d4"} Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.777825 4966 scope.go:117] "RemoveContainer" containerID="0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.778014 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x5qdp" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.787366 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-utilities\") pod \"bea1b22d-72c0-4558-845c-52ea15d2eb94\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.787436 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bghgk\" (UniqueName: \"kubernetes.io/projected/bea1b22d-72c0-4558-845c-52ea15d2eb94-kube-api-access-bghgk\") pod \"bea1b22d-72c0-4558-845c-52ea15d2eb94\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.787473 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-catalog-content\") pod \"bea1b22d-72c0-4558-845c-52ea15d2eb94\" (UID: \"bea1b22d-72c0-4558-845c-52ea15d2eb94\") " Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.788166 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-utilities" (OuterVolumeSpecName: "utilities") pod "bea1b22d-72c0-4558-845c-52ea15d2eb94" (UID: "bea1b22d-72c0-4558-845c-52ea15d2eb94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.792211 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.801027 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bea1b22d-72c0-4558-845c-52ea15d2eb94-kube-api-access-bghgk" (OuterVolumeSpecName: "kube-api-access-bghgk") pod "bea1b22d-72c0-4558-845c-52ea15d2eb94" (UID: "bea1b22d-72c0-4558-845c-52ea15d2eb94"). InnerVolumeSpecName "kube-api-access-bghgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.822855 4966 scope.go:117] "RemoveContainer" containerID="392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.853141 4966 scope.go:117] "RemoveContainer" containerID="10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.868742 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bea1b22d-72c0-4558-845c-52ea15d2eb94" (UID: "bea1b22d-72c0-4558-845c-52ea15d2eb94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.894188 4966 scope.go:117] "RemoveContainer" containerID="0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.894391 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bghgk\" (UniqueName: \"kubernetes.io/projected/bea1b22d-72c0-4558-845c-52ea15d2eb94-kube-api-access-bghgk\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.894445 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bea1b22d-72c0-4558-845c-52ea15d2eb94-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:23 crc kubenswrapper[4966]: E1217 08:58:23.896391 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa\": container with ID starting with 0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa not found: ID does not exist" containerID="0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.896597 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa"} err="failed to get container status \"0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa\": rpc error: code = NotFound desc = could not find container \"0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa\": container with ID starting with 0f70e5b2cb37fdd8ddc86b49fd277db2fc9285d1a80ffc47f0787b819dafe1aa not found: ID does not exist" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.896623 4966 scope.go:117] "RemoveContainer" containerID="392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8" Dec 17 08:58:23 crc kubenswrapper[4966]: E1217 08:58:23.897960 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8\": container with ID starting with 392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8 not found: ID does not exist" containerID="392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.897989 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8"} err="failed to get container status \"392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8\": rpc error: code = NotFound desc = could not find container \"392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8\": container with ID starting with 392ac6a1e64e8a23696e185a81df6291e7e45a23b7c758f8b9d04476a0d783e8 not found: ID does not exist" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.898005 4966 scope.go:117] "RemoveContainer" containerID="10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a" Dec 17 08:58:23 crc kubenswrapper[4966]: E1217 08:58:23.898298 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a\": container with ID starting with 10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a not found: ID does not exist" containerID="10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a" Dec 17 08:58:23 crc kubenswrapper[4966]: I1217 08:58:23.898328 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a"} err="failed to get container status \"10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a\": rpc error: code = NotFound desc = could not find container \"10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a\": container with ID starting with 10db1c37bee85a873a3b2bfc79e26533fc6b114a55d1fd975038995432108a6a not found: ID does not exist" Dec 17 08:58:24 crc kubenswrapper[4966]: I1217 08:58:24.115621 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x5qdp"] Dec 17 08:58:24 crc kubenswrapper[4966]: I1217 08:58:24.124200 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x5qdp"] Dec 17 08:58:24 crc kubenswrapper[4966]: I1217 08:58:24.860673 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bea1b22d-72c0-4558-845c-52ea15d2eb94" path="/var/lib/kubelet/pods/bea1b22d-72c0-4558-845c-52ea15d2eb94/volumes" Dec 17 08:58:54 crc kubenswrapper[4966]: I1217 08:58:54.055712 4966 generic.go:334] "Generic (PLEG): container finished" podID="c813431d-6629-4804-bf18-52c0f9634050" containerID="1b4aa098c6d1e2509350decdca7130996eed91d5ea387c7d035cc3ff67d85d43" exitCode=0 Dec 17 08:58:54 crc kubenswrapper[4966]: I1217 08:58:54.055933 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" event={"ID":"c813431d-6629-4804-bf18-52c0f9634050","Type":"ContainerDied","Data":"1b4aa098c6d1e2509350decdca7130996eed91d5ea387c7d035cc3ff67d85d43"} Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.542427 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.545945 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ssh-key\") pod \"c813431d-6629-4804-bf18-52c0f9634050\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.546041 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz4zt\" (UniqueName: \"kubernetes.io/projected/c813431d-6629-4804-bf18-52c0f9634050-kube-api-access-rz4zt\") pod \"c813431d-6629-4804-bf18-52c0f9634050\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.546206 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-inventory\") pod \"c813431d-6629-4804-bf18-52c0f9634050\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.546231 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ovn-combined-ca-bundle\") pod \"c813431d-6629-4804-bf18-52c0f9634050\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.546265 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c813431d-6629-4804-bf18-52c0f9634050-ovncontroller-config-0\") pod \"c813431d-6629-4804-bf18-52c0f9634050\" (UID: \"c813431d-6629-4804-bf18-52c0f9634050\") " Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.551254 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c813431d-6629-4804-bf18-52c0f9634050" (UID: "c813431d-6629-4804-bf18-52c0f9634050"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.557100 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c813431d-6629-4804-bf18-52c0f9634050-kube-api-access-rz4zt" (OuterVolumeSpecName: "kube-api-access-rz4zt") pod "c813431d-6629-4804-bf18-52c0f9634050" (UID: "c813431d-6629-4804-bf18-52c0f9634050"). InnerVolumeSpecName "kube-api-access-rz4zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.598275 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c813431d-6629-4804-bf18-52c0f9634050-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "c813431d-6629-4804-bf18-52c0f9634050" (UID: "c813431d-6629-4804-bf18-52c0f9634050"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.613644 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-inventory" (OuterVolumeSpecName: "inventory") pod "c813431d-6629-4804-bf18-52c0f9634050" (UID: "c813431d-6629-4804-bf18-52c0f9634050"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.614122 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c813431d-6629-4804-bf18-52c0f9634050" (UID: "c813431d-6629-4804-bf18-52c0f9634050"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.648005 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz4zt\" (UniqueName: \"kubernetes.io/projected/c813431d-6629-4804-bf18-52c0f9634050-kube-api-access-rz4zt\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.648034 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.648044 4966 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.648054 4966 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c813431d-6629-4804-bf18-52c0f9634050-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:55 crc kubenswrapper[4966]: I1217 08:58:55.648061 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c813431d-6629-4804-bf18-52c0f9634050-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.080573 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" event={"ID":"c813431d-6629-4804-bf18-52c0f9634050","Type":"ContainerDied","Data":"2a1c8952ef045db4b00edd473dbd0ec3afaded548b0655aa97458d364ef7ee81"} Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.080616 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a1c8952ef045db4b00edd473dbd0ec3afaded548b0655aa97458d364ef7ee81" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.080701 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-g5kkb" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.207655 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp"] Dec 17 08:58:56 crc kubenswrapper[4966]: E1217 08:58:56.208590 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerName="extract-utilities" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.208606 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerName="extract-utilities" Dec 17 08:58:56 crc kubenswrapper[4966]: E1217 08:58:56.208637 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerName="extract-content" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.208646 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerName="extract-content" Dec 17 08:58:56 crc kubenswrapper[4966]: E1217 08:58:56.208701 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerName="registry-server" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.208729 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerName="registry-server" Dec 17 08:58:56 crc kubenswrapper[4966]: E1217 08:58:56.208764 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c813431d-6629-4804-bf18-52c0f9634050" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.208772 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c813431d-6629-4804-bf18-52c0f9634050" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.209300 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c813431d-6629-4804-bf18-52c0f9634050" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.209326 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="bea1b22d-72c0-4558-845c-52ea15d2eb94" containerName="registry-server" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.212210 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.220574 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.220661 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.221735 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.223473 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.225010 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.238107 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.256664 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp"] Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.310761 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.310802 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.310822 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.310916 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rfgf\" (UniqueName: \"kubernetes.io/projected/46c3f44b-3d03-42c5-ab57-c8cf2405968b-kube-api-access-9rfgf\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.310942 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.311018 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.412685 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.412840 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.412893 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.412922 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.412962 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rfgf\" (UniqueName: \"kubernetes.io/projected/46c3f44b-3d03-42c5-ab57-c8cf2405968b-kube-api-access-9rfgf\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.412997 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.417132 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.417132 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.418200 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.418631 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.422605 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.429792 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rfgf\" (UniqueName: \"kubernetes.io/projected/46c3f44b-3d03-42c5-ab57-c8cf2405968b-kube-api-access-9rfgf\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:56 crc kubenswrapper[4966]: I1217 08:58:56.557548 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:58:57 crc kubenswrapper[4966]: I1217 08:58:57.084274 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp"] Dec 17 08:58:58 crc kubenswrapper[4966]: I1217 08:58:58.099298 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" event={"ID":"46c3f44b-3d03-42c5-ab57-c8cf2405968b","Type":"ContainerStarted","Data":"2d0240c578374e0972ae1c6fa19b95e262318dbf291317b53e0c2d52cfa3a4ed"} Dec 17 08:58:58 crc kubenswrapper[4966]: I1217 08:58:58.099531 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" event={"ID":"46c3f44b-3d03-42c5-ab57-c8cf2405968b","Type":"ContainerStarted","Data":"08cd6e3e81585cd3506f848cd6784db4770de757e9ba890aeced0d973f88e203"} Dec 17 08:58:58 crc kubenswrapper[4966]: I1217 08:58:58.123712 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" podStartSLOduration=1.5587612640000001 podStartE2EDuration="2.123691243s" podCreationTimestamp="2025-12-17 08:58:56 +0000 UTC" firstStartedPulling="2025-12-17 08:58:57.101362538 +0000 UTC m=+2272.646432480" lastFinishedPulling="2025-12-17 08:58:57.666292517 +0000 UTC m=+2273.211362459" observedRunningTime="2025-12-17 08:58:58.1144301 +0000 UTC m=+2273.659500042" watchObservedRunningTime="2025-12-17 08:58:58.123691243 +0000 UTC m=+2273.668761185" Dec 17 08:59:54 crc kubenswrapper[4966]: I1217 08:59:54.628533 4966 generic.go:334] "Generic (PLEG): container finished" podID="46c3f44b-3d03-42c5-ab57-c8cf2405968b" containerID="2d0240c578374e0972ae1c6fa19b95e262318dbf291317b53e0c2d52cfa3a4ed" exitCode=0 Dec 17 08:59:54 crc kubenswrapper[4966]: I1217 08:59:54.628613 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" event={"ID":"46c3f44b-3d03-42c5-ab57-c8cf2405968b","Type":"ContainerDied","Data":"2d0240c578374e0972ae1c6fa19b95e262318dbf291317b53e0c2d52cfa3a4ed"} Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.173751 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.190607 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-inventory\") pod \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.190701 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-ssh-key\") pod \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.262439 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-inventory" (OuterVolumeSpecName: "inventory") pod "46c3f44b-3d03-42c5-ab57-c8cf2405968b" (UID: "46c3f44b-3d03-42c5-ab57-c8cf2405968b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.269445 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "46c3f44b-3d03-42c5-ab57-c8cf2405968b" (UID: "46c3f44b-3d03-42c5-ab57-c8cf2405968b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.293256 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.294245 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rfgf\" (UniqueName: \"kubernetes.io/projected/46c3f44b-3d03-42c5-ab57-c8cf2405968b-kube-api-access-9rfgf\") pod \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.294348 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-nova-metadata-neutron-config-0\") pod \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.294420 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-metadata-combined-ca-bundle\") pod \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\" (UID: \"46c3f44b-3d03-42c5-ab57-c8cf2405968b\") " Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.295232 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.295256 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.298366 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "46c3f44b-3d03-42c5-ab57-c8cf2405968b" (UID: "46c3f44b-3d03-42c5-ab57-c8cf2405968b"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.298498 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46c3f44b-3d03-42c5-ab57-c8cf2405968b-kube-api-access-9rfgf" (OuterVolumeSpecName: "kube-api-access-9rfgf") pod "46c3f44b-3d03-42c5-ab57-c8cf2405968b" (UID: "46c3f44b-3d03-42c5-ab57-c8cf2405968b"). InnerVolumeSpecName "kube-api-access-9rfgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.317081 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "46c3f44b-3d03-42c5-ab57-c8cf2405968b" (UID: "46c3f44b-3d03-42c5-ab57-c8cf2405968b"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.336669 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "46c3f44b-3d03-42c5-ab57-c8cf2405968b" (UID: "46c3f44b-3d03-42c5-ab57-c8cf2405968b"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.395905 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rfgf\" (UniqueName: \"kubernetes.io/projected/46c3f44b-3d03-42c5-ab57-c8cf2405968b-kube-api-access-9rfgf\") on node \"crc\" DevicePath \"\"" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.395936 4966 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.395947 4966 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.395960 4966 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/46c3f44b-3d03-42c5-ab57-c8cf2405968b-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.656995 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" event={"ID":"46c3f44b-3d03-42c5-ab57-c8cf2405968b","Type":"ContainerDied","Data":"08cd6e3e81585cd3506f848cd6784db4770de757e9ba890aeced0d973f88e203"} Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.657046 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08cd6e3e81585cd3506f848cd6784db4770de757e9ba890aeced0d973f88e203" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.657120 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cvlkp" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.867950 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c"] Dec 17 08:59:56 crc kubenswrapper[4966]: E1217 08:59:56.868797 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c3f44b-3d03-42c5-ab57-c8cf2405968b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.868818 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c3f44b-3d03-42c5-ab57-c8cf2405968b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.869088 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="46c3f44b-3d03-42c5-ab57-c8cf2405968b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.870155 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.874991 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.875065 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.875219 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.875412 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.875442 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.886331 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c"] Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.927964 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.928002 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9qmr\" (UniqueName: \"kubernetes.io/projected/02dd574e-7cc4-4736-9772-94abd7f058f9-kube-api-access-t9qmr\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.928064 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.928144 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:56 crc kubenswrapper[4966]: I1217 08:59:56.928282 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.029570 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.029631 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9qmr\" (UniqueName: \"kubernetes.io/projected/02dd574e-7cc4-4736-9772-94abd7f058f9-kube-api-access-t9qmr\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.029660 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.029692 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.029732 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.034014 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.034200 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.034458 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.037200 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.049911 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9qmr\" (UniqueName: \"kubernetes.io/projected/02dd574e-7cc4-4736-9772-94abd7f058f9-kube-api-access-t9qmr\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.248864 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 08:59:57 crc kubenswrapper[4966]: I1217 08:59:57.862967 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c"] Dec 17 08:59:58 crc kubenswrapper[4966]: I1217 08:59:58.679385 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" event={"ID":"02dd574e-7cc4-4736-9772-94abd7f058f9","Type":"ContainerStarted","Data":"7b2242250cc43c8da169d18988449699196d3e89c787fb4542648db8589ebc84"} Dec 17 08:59:59 crc kubenswrapper[4966]: I1217 08:59:59.690553 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" event={"ID":"02dd574e-7cc4-4736-9772-94abd7f058f9","Type":"ContainerStarted","Data":"38851e539fc0826cd84c33b7453504aa4bd13de7802fdea23a2c26835e53583e"} Dec 17 08:59:59 crc kubenswrapper[4966]: I1217 08:59:59.713652 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" podStartSLOduration=2.86524617 podStartE2EDuration="3.71362574s" podCreationTimestamp="2025-12-17 08:59:56 +0000 UTC" firstStartedPulling="2025-12-17 08:59:57.870910329 +0000 UTC m=+2333.415980271" lastFinishedPulling="2025-12-17 08:59:58.719289909 +0000 UTC m=+2334.264359841" observedRunningTime="2025-12-17 08:59:59.707954995 +0000 UTC m=+2335.253024967" watchObservedRunningTime="2025-12-17 08:59:59.71362574 +0000 UTC m=+2335.258695722" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.141432 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs"] Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.143569 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.149534 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.149906 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.159263 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs"] Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.294571 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-config-volume\") pod \"collect-profiles-29432700-f8ffs\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.294679 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr5ff\" (UniqueName: \"kubernetes.io/projected/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-kube-api-access-zr5ff\") pod \"collect-profiles-29432700-f8ffs\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.295482 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-secret-volume\") pod \"collect-profiles-29432700-f8ffs\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.397663 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-config-volume\") pod \"collect-profiles-29432700-f8ffs\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.397729 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr5ff\" (UniqueName: \"kubernetes.io/projected/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-kube-api-access-zr5ff\") pod \"collect-profiles-29432700-f8ffs\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.397781 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-secret-volume\") pod \"collect-profiles-29432700-f8ffs\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.399067 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-config-volume\") pod \"collect-profiles-29432700-f8ffs\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.402748 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-secret-volume\") pod \"collect-profiles-29432700-f8ffs\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.428631 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr5ff\" (UniqueName: \"kubernetes.io/projected/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-kube-api-access-zr5ff\") pod \"collect-profiles-29432700-f8ffs\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.471355 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:00 crc kubenswrapper[4966]: I1217 09:00:00.961192 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs"] Dec 17 09:00:01 crc kubenswrapper[4966]: I1217 09:00:01.708017 4966 generic.go:334] "Generic (PLEG): container finished" podID="84d0cb6a-efa1-4c3a-bd5e-565685eeea80" containerID="d7e1c0bd4a32d96280b9f82f2a31ead45b81dc91be7e7afbe8f14844757a3543" exitCode=0 Dec 17 09:00:01 crc kubenswrapper[4966]: I1217 09:00:01.708134 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" event={"ID":"84d0cb6a-efa1-4c3a-bd5e-565685eeea80","Type":"ContainerDied","Data":"d7e1c0bd4a32d96280b9f82f2a31ead45b81dc91be7e7afbe8f14844757a3543"} Dec 17 09:00:01 crc kubenswrapper[4966]: I1217 09:00:01.708512 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" event={"ID":"84d0cb6a-efa1-4c3a-bd5e-565685eeea80","Type":"ContainerStarted","Data":"0a131f88891276b02fb87aeab4c61fb5bbab86da2278270f0c9e438cd99ef540"} Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.070372 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.156709 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr5ff\" (UniqueName: \"kubernetes.io/projected/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-kube-api-access-zr5ff\") pod \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.156781 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-secret-volume\") pod \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.156855 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-config-volume\") pod \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\" (UID: \"84d0cb6a-efa1-4c3a-bd5e-565685eeea80\") " Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.157648 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-config-volume" (OuterVolumeSpecName: "config-volume") pod "84d0cb6a-efa1-4c3a-bd5e-565685eeea80" (UID: "84d0cb6a-efa1-4c3a-bd5e-565685eeea80"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.162936 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "84d0cb6a-efa1-4c3a-bd5e-565685eeea80" (UID: "84d0cb6a-efa1-4c3a-bd5e-565685eeea80"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.163001 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-kube-api-access-zr5ff" (OuterVolumeSpecName: "kube-api-access-zr5ff") pod "84d0cb6a-efa1-4c3a-bd5e-565685eeea80" (UID: "84d0cb6a-efa1-4c3a-bd5e-565685eeea80"). InnerVolumeSpecName "kube-api-access-zr5ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.258205 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr5ff\" (UniqueName: \"kubernetes.io/projected/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-kube-api-access-zr5ff\") on node \"crc\" DevicePath \"\"" Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.258232 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.258242 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84d0cb6a-efa1-4c3a-bd5e-565685eeea80-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.735075 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" event={"ID":"84d0cb6a-efa1-4c3a-bd5e-565685eeea80","Type":"ContainerDied","Data":"0a131f88891276b02fb87aeab4c61fb5bbab86da2278270f0c9e438cd99ef540"} Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.735114 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a131f88891276b02fb87aeab4c61fb5bbab86da2278270f0c9e438cd99ef540" Dec 17 09:00:03 crc kubenswrapper[4966]: I1217 09:00:03.735169 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs" Dec 17 09:00:04 crc kubenswrapper[4966]: I1217 09:00:04.187741 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4"] Dec 17 09:00:04 crc kubenswrapper[4966]: I1217 09:00:04.203817 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432655-x4nw4"] Dec 17 09:00:04 crc kubenswrapper[4966]: I1217 09:00:04.843660 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bc07aea-e7b5-473c-957c-a5295edc2f4e" path="/var/lib/kubelet/pods/9bc07aea-e7b5-473c-957c-a5295edc2f4e/volumes" Dec 17 09:00:16 crc kubenswrapper[4966]: I1217 09:00:16.807918 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:00:16 crc kubenswrapper[4966]: I1217 09:00:16.808413 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:00:46 crc kubenswrapper[4966]: I1217 09:00:46.807705 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:00:46 crc kubenswrapper[4966]: I1217 09:00:46.808508 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:00:51 crc kubenswrapper[4966]: I1217 09:00:51.303110 4966 scope.go:117] "RemoveContainer" containerID="5057125f0113eedefe674fe8786bd1bfb196007e3dc353d61e0172c0bfe86551" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.152171 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29432701-k9fm6"] Dec 17 09:01:00 crc kubenswrapper[4966]: E1217 09:01:00.154390 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d0cb6a-efa1-4c3a-bd5e-565685eeea80" containerName="collect-profiles" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.154415 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d0cb6a-efa1-4c3a-bd5e-565685eeea80" containerName="collect-profiles" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.154703 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d0cb6a-efa1-4c3a-bd5e-565685eeea80" containerName="collect-profiles" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.155596 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.171457 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29432701-k9fm6"] Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.220111 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-fernet-keys\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.220249 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-combined-ca-bundle\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.220297 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd9ls\" (UniqueName: \"kubernetes.io/projected/bd93c94b-5390-494f-9626-77603826668e-kube-api-access-dd9ls\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.220399 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-config-data\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.321825 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-fernet-keys\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.321990 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-combined-ca-bundle\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.322031 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd9ls\" (UniqueName: \"kubernetes.io/projected/bd93c94b-5390-494f-9626-77603826668e-kube-api-access-dd9ls\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.322132 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-config-data\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.330986 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-fernet-keys\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.331353 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-combined-ca-bundle\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.349972 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd9ls\" (UniqueName: \"kubernetes.io/projected/bd93c94b-5390-494f-9626-77603826668e-kube-api-access-dd9ls\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.351902 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-config-data\") pod \"keystone-cron-29432701-k9fm6\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:00 crc kubenswrapper[4966]: I1217 09:01:00.529542 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:01 crc kubenswrapper[4966]: I1217 09:01:01.045130 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29432701-k9fm6"] Dec 17 09:01:01 crc kubenswrapper[4966]: I1217 09:01:01.444863 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432701-k9fm6" event={"ID":"bd93c94b-5390-494f-9626-77603826668e","Type":"ContainerStarted","Data":"36fc036b2b3e41280672b83fcf44d12dd20d339b5163dc110e505ed80adeb761"} Dec 17 09:01:01 crc kubenswrapper[4966]: I1217 09:01:01.444933 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432701-k9fm6" event={"ID":"bd93c94b-5390-494f-9626-77603826668e","Type":"ContainerStarted","Data":"3c46afb578f709032010ddc7c110d12e6ea39329a4dec50c41334180ea5dfedc"} Dec 17 09:01:01 crc kubenswrapper[4966]: I1217 09:01:01.477555 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29432701-k9fm6" podStartSLOduration=1.477530274 podStartE2EDuration="1.477530274s" podCreationTimestamp="2025-12-17 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 09:01:01.470455541 +0000 UTC m=+2397.015525483" watchObservedRunningTime="2025-12-17 09:01:01.477530274 +0000 UTC m=+2397.022600226" Dec 17 09:01:04 crc kubenswrapper[4966]: I1217 09:01:04.473986 4966 generic.go:334] "Generic (PLEG): container finished" podID="bd93c94b-5390-494f-9626-77603826668e" containerID="36fc036b2b3e41280672b83fcf44d12dd20d339b5163dc110e505ed80adeb761" exitCode=0 Dec 17 09:01:04 crc kubenswrapper[4966]: I1217 09:01:04.474058 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432701-k9fm6" event={"ID":"bd93c94b-5390-494f-9626-77603826668e","Type":"ContainerDied","Data":"36fc036b2b3e41280672b83fcf44d12dd20d339b5163dc110e505ed80adeb761"} Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.817472 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.842959 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-config-data\") pod \"bd93c94b-5390-494f-9626-77603826668e\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.843111 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-combined-ca-bundle\") pod \"bd93c94b-5390-494f-9626-77603826668e\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.843165 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-fernet-keys\") pod \"bd93c94b-5390-494f-9626-77603826668e\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.843321 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd9ls\" (UniqueName: \"kubernetes.io/projected/bd93c94b-5390-494f-9626-77603826668e-kube-api-access-dd9ls\") pod \"bd93c94b-5390-494f-9626-77603826668e\" (UID: \"bd93c94b-5390-494f-9626-77603826668e\") " Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.850247 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bd93c94b-5390-494f-9626-77603826668e" (UID: "bd93c94b-5390-494f-9626-77603826668e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.865057 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd93c94b-5390-494f-9626-77603826668e-kube-api-access-dd9ls" (OuterVolumeSpecName: "kube-api-access-dd9ls") pod "bd93c94b-5390-494f-9626-77603826668e" (UID: "bd93c94b-5390-494f-9626-77603826668e"). InnerVolumeSpecName "kube-api-access-dd9ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.886827 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd93c94b-5390-494f-9626-77603826668e" (UID: "bd93c94b-5390-494f-9626-77603826668e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.921472 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-config-data" (OuterVolumeSpecName: "config-data") pod "bd93c94b-5390-494f-9626-77603826668e" (UID: "bd93c94b-5390-494f-9626-77603826668e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.947239 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd9ls\" (UniqueName: \"kubernetes.io/projected/bd93c94b-5390-494f-9626-77603826668e-kube-api-access-dd9ls\") on node \"crc\" DevicePath \"\"" Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.947492 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.947607 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 09:01:05 crc kubenswrapper[4966]: I1217 09:01:05.947667 4966 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bd93c94b-5390-494f-9626-77603826668e-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 17 09:01:06 crc kubenswrapper[4966]: I1217 09:01:06.493594 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432701-k9fm6" event={"ID":"bd93c94b-5390-494f-9626-77603826668e","Type":"ContainerDied","Data":"3c46afb578f709032010ddc7c110d12e6ea39329a4dec50c41334180ea5dfedc"} Dec 17 09:01:06 crc kubenswrapper[4966]: I1217 09:01:06.493638 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c46afb578f709032010ddc7c110d12e6ea39329a4dec50c41334180ea5dfedc" Dec 17 09:01:06 crc kubenswrapper[4966]: I1217 09:01:06.493679 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432701-k9fm6" Dec 17 09:01:16 crc kubenswrapper[4966]: I1217 09:01:16.808407 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:01:16 crc kubenswrapper[4966]: I1217 09:01:16.809135 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:01:16 crc kubenswrapper[4966]: I1217 09:01:16.809198 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:01:16 crc kubenswrapper[4966]: I1217 09:01:16.810238 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:01:16 crc kubenswrapper[4966]: I1217 09:01:16.810345 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" gracePeriod=600 Dec 17 09:01:16 crc kubenswrapper[4966]: E1217 09:01:16.935388 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:01:17 crc kubenswrapper[4966]: I1217 09:01:17.613172 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" exitCode=0 Dec 17 09:01:17 crc kubenswrapper[4966]: I1217 09:01:17.613498 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592"} Dec 17 09:01:17 crc kubenswrapper[4966]: I1217 09:01:17.613538 4966 scope.go:117] "RemoveContainer" containerID="4757810d46cf706e3c032dafdc0e43a6e21fb930b0834212015c013294922d1f" Dec 17 09:01:17 crc kubenswrapper[4966]: I1217 09:01:17.614332 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:01:17 crc kubenswrapper[4966]: E1217 09:01:17.614704 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:01:31 crc kubenswrapper[4966]: I1217 09:01:31.832530 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:01:31 crc kubenswrapper[4966]: E1217 09:01:31.833221 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:01:45 crc kubenswrapper[4966]: I1217 09:01:45.831164 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:01:45 crc kubenswrapper[4966]: E1217 09:01:45.832387 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:01:56 crc kubenswrapper[4966]: I1217 09:01:56.830977 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:01:56 crc kubenswrapper[4966]: E1217 09:01:56.831773 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:02:08 crc kubenswrapper[4966]: I1217 09:02:08.831045 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:02:08 crc kubenswrapper[4966]: E1217 09:02:08.832284 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:02:19 crc kubenswrapper[4966]: I1217 09:02:19.830912 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:02:19 crc kubenswrapper[4966]: E1217 09:02:19.831598 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:02:31 crc kubenswrapper[4966]: I1217 09:02:31.830550 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:02:31 crc kubenswrapper[4966]: E1217 09:02:31.831466 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:02:43 crc kubenswrapper[4966]: I1217 09:02:43.831373 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:02:43 crc kubenswrapper[4966]: E1217 09:02:43.832203 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:02:58 crc kubenswrapper[4966]: I1217 09:02:58.830697 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:02:58 crc kubenswrapper[4966]: E1217 09:02:58.832553 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:03:10 crc kubenswrapper[4966]: I1217 09:03:10.830946 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:03:10 crc kubenswrapper[4966]: E1217 09:03:10.831652 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:03:24 crc kubenswrapper[4966]: I1217 09:03:24.837654 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:03:24 crc kubenswrapper[4966]: E1217 09:03:24.838509 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:03:38 crc kubenswrapper[4966]: I1217 09:03:38.830508 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:03:38 crc kubenswrapper[4966]: E1217 09:03:38.831314 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:03:49 crc kubenswrapper[4966]: I1217 09:03:49.830311 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:03:49 crc kubenswrapper[4966]: E1217 09:03:49.831460 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:04:01 crc kubenswrapper[4966]: I1217 09:04:01.831356 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:04:01 crc kubenswrapper[4966]: E1217 09:04:01.832447 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:04:16 crc kubenswrapper[4966]: I1217 09:04:16.831274 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:04:16 crc kubenswrapper[4966]: E1217 09:04:16.831932 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:04:24 crc kubenswrapper[4966]: I1217 09:04:24.383682 4966 generic.go:334] "Generic (PLEG): container finished" podID="02dd574e-7cc4-4736-9772-94abd7f058f9" containerID="38851e539fc0826cd84c33b7453504aa4bd13de7802fdea23a2c26835e53583e" exitCode=0 Dec 17 09:04:24 crc kubenswrapper[4966]: I1217 09:04:24.383775 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" event={"ID":"02dd574e-7cc4-4736-9772-94abd7f058f9","Type":"ContainerDied","Data":"38851e539fc0826cd84c33b7453504aa4bd13de7802fdea23a2c26835e53583e"} Dec 17 09:04:25 crc kubenswrapper[4966]: I1217 09:04:25.934169 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.052513 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-inventory\") pod \"02dd574e-7cc4-4736-9772-94abd7f058f9\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.052585 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-ssh-key\") pod \"02dd574e-7cc4-4736-9772-94abd7f058f9\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.052727 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-combined-ca-bundle\") pod \"02dd574e-7cc4-4736-9772-94abd7f058f9\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.052754 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9qmr\" (UniqueName: \"kubernetes.io/projected/02dd574e-7cc4-4736-9772-94abd7f058f9-kube-api-access-t9qmr\") pod \"02dd574e-7cc4-4736-9772-94abd7f058f9\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.052857 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-secret-0\") pod \"02dd574e-7cc4-4736-9772-94abd7f058f9\" (UID: \"02dd574e-7cc4-4736-9772-94abd7f058f9\") " Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.091210 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02dd574e-7cc4-4736-9772-94abd7f058f9-kube-api-access-t9qmr" (OuterVolumeSpecName: "kube-api-access-t9qmr") pod "02dd574e-7cc4-4736-9772-94abd7f058f9" (UID: "02dd574e-7cc4-4736-9772-94abd7f058f9"). InnerVolumeSpecName "kube-api-access-t9qmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.097483 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "02dd574e-7cc4-4736-9772-94abd7f058f9" (UID: "02dd574e-7cc4-4736-9772-94abd7f058f9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.131730 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-inventory" (OuterVolumeSpecName: "inventory") pod "02dd574e-7cc4-4736-9772-94abd7f058f9" (UID: "02dd574e-7cc4-4736-9772-94abd7f058f9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.135004 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "02dd574e-7cc4-4736-9772-94abd7f058f9" (UID: "02dd574e-7cc4-4736-9772-94abd7f058f9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.137311 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "02dd574e-7cc4-4736-9772-94abd7f058f9" (UID: "02dd574e-7cc4-4736-9772-94abd7f058f9"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.154972 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.155003 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.155016 4966 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.155041 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9qmr\" (UniqueName: \"kubernetes.io/projected/02dd574e-7cc4-4736-9772-94abd7f058f9-kube-api-access-t9qmr\") on node \"crc\" DevicePath \"\"" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.155052 4966 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/02dd574e-7cc4-4736-9772-94abd7f058f9-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.406414 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" event={"ID":"02dd574e-7cc4-4736-9772-94abd7f058f9","Type":"ContainerDied","Data":"7b2242250cc43c8da169d18988449699196d3e89c787fb4542648db8589ebc84"} Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.406463 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b2242250cc43c8da169d18988449699196d3e89c787fb4542648db8589ebc84" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.406476 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-v6b9c" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.500917 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7"] Dec 17 09:04:26 crc kubenswrapper[4966]: E1217 09:04:26.503328 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02dd574e-7cc4-4736-9772-94abd7f058f9" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.503360 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="02dd574e-7cc4-4736-9772-94abd7f058f9" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 17 09:04:26 crc kubenswrapper[4966]: E1217 09:04:26.503394 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd93c94b-5390-494f-9626-77603826668e" containerName="keystone-cron" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.503402 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd93c94b-5390-494f-9626-77603826668e" containerName="keystone-cron" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.504189 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="02dd574e-7cc4-4736-9772-94abd7f058f9" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.504231 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd93c94b-5390-494f-9626-77603826668e" containerName="keystone-cron" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.505007 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.509898 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.510219 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.510331 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.510443 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.510558 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.511064 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.514321 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.516139 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7"] Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.561243 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.561300 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.561339 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.561359 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.561412 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.561486 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hds4n\" (UniqueName: \"kubernetes.io/projected/c4bac84a-1f28-496b-91f2-c650fac7ed7d-kube-api-access-hds4n\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.561510 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.561559 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.561645 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.663988 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.664040 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.664154 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.664184 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.664212 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.664231 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.664258 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.664294 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hds4n\" (UniqueName: \"kubernetes.io/projected/c4bac84a-1f28-496b-91f2-c650fac7ed7d-kube-api-access-hds4n\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.664318 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.664989 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.673245 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.673909 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.674172 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.675546 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.675602 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.678809 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.679939 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.686405 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hds4n\" (UniqueName: \"kubernetes.io/projected/c4bac84a-1f28-496b-91f2-c650fac7ed7d-kube-api-access-hds4n\") pod \"nova-edpm-deployment-openstack-edpm-ipam-gf4m7\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:26 crc kubenswrapper[4966]: I1217 09:04:26.824906 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:04:27 crc kubenswrapper[4966]: I1217 09:04:27.402864 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7"] Dec 17 09:04:27 crc kubenswrapper[4966]: W1217 09:04:27.409235 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4bac84a_1f28_496b_91f2_c650fac7ed7d.slice/crio-9f4f0eba91182c175927aa4c8e2ad64864c0668c848279d23ae979f62a5f032d WatchSource:0}: Error finding container 9f4f0eba91182c175927aa4c8e2ad64864c0668c848279d23ae979f62a5f032d: Status 404 returned error can't find the container with id 9f4f0eba91182c175927aa4c8e2ad64864c0668c848279d23ae979f62a5f032d Dec 17 09:04:27 crc kubenswrapper[4966]: I1217 09:04:27.412685 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 09:04:28 crc kubenswrapper[4966]: I1217 09:04:28.432641 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" event={"ID":"c4bac84a-1f28-496b-91f2-c650fac7ed7d","Type":"ContainerStarted","Data":"9f4f0eba91182c175927aa4c8e2ad64864c0668c848279d23ae979f62a5f032d"} Dec 17 09:04:28 crc kubenswrapper[4966]: I1217 09:04:28.831209 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:04:28 crc kubenswrapper[4966]: E1217 09:04:28.831469 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:04:29 crc kubenswrapper[4966]: I1217 09:04:29.445565 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" event={"ID":"c4bac84a-1f28-496b-91f2-c650fac7ed7d","Type":"ContainerStarted","Data":"723aeb1f2fa584a08b7547d4df3bda9b7920144df5a6b588b93ec444e43fd982"} Dec 17 09:04:29 crc kubenswrapper[4966]: I1217 09:04:29.475150 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" podStartSLOduration=1.829554022 podStartE2EDuration="3.475131242s" podCreationTimestamp="2025-12-17 09:04:26 +0000 UTC" firstStartedPulling="2025-12-17 09:04:27.412369914 +0000 UTC m=+2602.957439856" lastFinishedPulling="2025-12-17 09:04:29.057947134 +0000 UTC m=+2604.603017076" observedRunningTime="2025-12-17 09:04:29.462364923 +0000 UTC m=+2605.007434865" watchObservedRunningTime="2025-12-17 09:04:29.475131242 +0000 UTC m=+2605.020201184" Dec 17 09:04:43 crc kubenswrapper[4966]: I1217 09:04:43.831179 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:04:43 crc kubenswrapper[4966]: E1217 09:04:43.832260 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:04:56 crc kubenswrapper[4966]: I1217 09:04:56.830377 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:04:56 crc kubenswrapper[4966]: E1217 09:04:56.831255 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:05:07 crc kubenswrapper[4966]: I1217 09:05:07.831810 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:05:07 crc kubenswrapper[4966]: E1217 09:05:07.833032 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:05:21 crc kubenswrapper[4966]: I1217 09:05:21.830621 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:05:21 crc kubenswrapper[4966]: E1217 09:05:21.831563 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:05:35 crc kubenswrapper[4966]: I1217 09:05:35.830409 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:05:35 crc kubenswrapper[4966]: E1217 09:05:35.831159 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:05:47 crc kubenswrapper[4966]: I1217 09:05:47.831354 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:05:47 crc kubenswrapper[4966]: E1217 09:05:47.832759 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:06:01 crc kubenswrapper[4966]: I1217 09:06:01.831012 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:06:01 crc kubenswrapper[4966]: E1217 09:06:01.831857 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:06:12 crc kubenswrapper[4966]: I1217 09:06:12.831710 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:06:12 crc kubenswrapper[4966]: E1217 09:06:12.832519 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:06:25 crc kubenswrapper[4966]: I1217 09:06:25.831972 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:06:26 crc kubenswrapper[4966]: I1217 09:06:26.653968 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"246ff596bfcf475a67b9d725efaf8f2795c418a5f24501a9f96ed1f28d380e19"} Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.638453 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nrsr2"] Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.641043 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.650349 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nrsr2"] Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.686748 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-utilities\") pod \"redhat-operators-nrsr2\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.686813 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xsp\" (UniqueName: \"kubernetes.io/projected/0a571f6a-6550-44e6-8cc7-17d01769d68a-kube-api-access-b7xsp\") pod \"redhat-operators-nrsr2\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.686917 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-catalog-content\") pod \"redhat-operators-nrsr2\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.788362 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xsp\" (UniqueName: \"kubernetes.io/projected/0a571f6a-6550-44e6-8cc7-17d01769d68a-kube-api-access-b7xsp\") pod \"redhat-operators-nrsr2\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.789214 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-catalog-content\") pod \"redhat-operators-nrsr2\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.789598 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-utilities\") pod \"redhat-operators-nrsr2\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.789703 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-catalog-content\") pod \"redhat-operators-nrsr2\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.790018 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-utilities\") pod \"redhat-operators-nrsr2\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:55 crc kubenswrapper[4966]: I1217 09:06:55.808843 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xsp\" (UniqueName: \"kubernetes.io/projected/0a571f6a-6550-44e6-8cc7-17d01769d68a-kube-api-access-b7xsp\") pod \"redhat-operators-nrsr2\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:56 crc kubenswrapper[4966]: I1217 09:06:56.005781 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:06:56 crc kubenswrapper[4966]: I1217 09:06:56.573643 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nrsr2"] Dec 17 09:06:56 crc kubenswrapper[4966]: I1217 09:06:56.938130 4966 generic.go:334] "Generic (PLEG): container finished" podID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerID="1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779" exitCode=0 Dec 17 09:06:56 crc kubenswrapper[4966]: I1217 09:06:56.938247 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrsr2" event={"ID":"0a571f6a-6550-44e6-8cc7-17d01769d68a","Type":"ContainerDied","Data":"1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779"} Dec 17 09:06:56 crc kubenswrapper[4966]: I1217 09:06:56.938455 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrsr2" event={"ID":"0a571f6a-6550-44e6-8cc7-17d01769d68a","Type":"ContainerStarted","Data":"5209302b2a31e64438947dac9e7321116d97814420ae8292257deda5454fce3b"} Dec 17 09:06:59 crc kubenswrapper[4966]: I1217 09:06:59.965915 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrsr2" event={"ID":"0a571f6a-6550-44e6-8cc7-17d01769d68a","Type":"ContainerStarted","Data":"c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa"} Dec 17 09:07:04 crc kubenswrapper[4966]: I1217 09:07:04.010004 4966 generic.go:334] "Generic (PLEG): container finished" podID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerID="c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa" exitCode=0 Dec 17 09:07:04 crc kubenswrapper[4966]: I1217 09:07:04.010288 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrsr2" event={"ID":"0a571f6a-6550-44e6-8cc7-17d01769d68a","Type":"ContainerDied","Data":"c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa"} Dec 17 09:07:05 crc kubenswrapper[4966]: I1217 09:07:05.023231 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrsr2" event={"ID":"0a571f6a-6550-44e6-8cc7-17d01769d68a","Type":"ContainerStarted","Data":"6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb"} Dec 17 09:07:05 crc kubenswrapper[4966]: I1217 09:07:05.046986 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nrsr2" podStartSLOduration=2.407088799 podStartE2EDuration="10.04696969s" podCreationTimestamp="2025-12-17 09:06:55 +0000 UTC" firstStartedPulling="2025-12-17 09:06:56.940321069 +0000 UTC m=+2752.485391011" lastFinishedPulling="2025-12-17 09:07:04.58020197 +0000 UTC m=+2760.125271902" observedRunningTime="2025-12-17 09:07:05.039464695 +0000 UTC m=+2760.584534657" watchObservedRunningTime="2025-12-17 09:07:05.04696969 +0000 UTC m=+2760.592039632" Dec 17 09:07:06 crc kubenswrapper[4966]: I1217 09:07:06.006599 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:07:06 crc kubenswrapper[4966]: I1217 09:07:06.006942 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:07:07 crc kubenswrapper[4966]: I1217 09:07:07.055409 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nrsr2" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="registry-server" probeResult="failure" output=< Dec 17 09:07:07 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:07:07 crc kubenswrapper[4966]: > Dec 17 09:07:17 crc kubenswrapper[4966]: I1217 09:07:17.053733 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nrsr2" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="registry-server" probeResult="failure" output=< Dec 17 09:07:17 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:07:17 crc kubenswrapper[4966]: > Dec 17 09:07:26 crc kubenswrapper[4966]: I1217 09:07:26.059063 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:07:26 crc kubenswrapper[4966]: I1217 09:07:26.113673 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:07:26 crc kubenswrapper[4966]: I1217 09:07:26.825280 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nrsr2"] Dec 17 09:07:27 crc kubenswrapper[4966]: I1217 09:07:27.236052 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nrsr2" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="registry-server" containerID="cri-o://6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb" gracePeriod=2 Dec 17 09:07:27 crc kubenswrapper[4966]: I1217 09:07:27.772404 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:07:27 crc kubenswrapper[4966]: I1217 09:07:27.864272 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7xsp\" (UniqueName: \"kubernetes.io/projected/0a571f6a-6550-44e6-8cc7-17d01769d68a-kube-api-access-b7xsp\") pod \"0a571f6a-6550-44e6-8cc7-17d01769d68a\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " Dec 17 09:07:27 crc kubenswrapper[4966]: I1217 09:07:27.864351 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-catalog-content\") pod \"0a571f6a-6550-44e6-8cc7-17d01769d68a\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " Dec 17 09:07:27 crc kubenswrapper[4966]: I1217 09:07:27.864563 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-utilities\") pod \"0a571f6a-6550-44e6-8cc7-17d01769d68a\" (UID: \"0a571f6a-6550-44e6-8cc7-17d01769d68a\") " Dec 17 09:07:27 crc kubenswrapper[4966]: I1217 09:07:27.865544 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-utilities" (OuterVolumeSpecName: "utilities") pod "0a571f6a-6550-44e6-8cc7-17d01769d68a" (UID: "0a571f6a-6550-44e6-8cc7-17d01769d68a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:07:27 crc kubenswrapper[4966]: I1217 09:07:27.915094 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a571f6a-6550-44e6-8cc7-17d01769d68a-kube-api-access-b7xsp" (OuterVolumeSpecName: "kube-api-access-b7xsp") pod "0a571f6a-6550-44e6-8cc7-17d01769d68a" (UID: "0a571f6a-6550-44e6-8cc7-17d01769d68a"). InnerVolumeSpecName "kube-api-access-b7xsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:07:27 crc kubenswrapper[4966]: I1217 09:07:27.967830 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:27 crc kubenswrapper[4966]: I1217 09:07:27.967857 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7xsp\" (UniqueName: \"kubernetes.io/projected/0a571f6a-6550-44e6-8cc7-17d01769d68a-kube-api-access-b7xsp\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.038749 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a571f6a-6550-44e6-8cc7-17d01769d68a" (UID: "0a571f6a-6550-44e6-8cc7-17d01769d68a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.070444 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a571f6a-6550-44e6-8cc7-17d01769d68a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.246099 4966 generic.go:334] "Generic (PLEG): container finished" podID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerID="6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb" exitCode=0 Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.246156 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrsr2" event={"ID":"0a571f6a-6550-44e6-8cc7-17d01769d68a","Type":"ContainerDied","Data":"6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb"} Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.246167 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrsr2" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.246188 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrsr2" event={"ID":"0a571f6a-6550-44e6-8cc7-17d01769d68a","Type":"ContainerDied","Data":"5209302b2a31e64438947dac9e7321116d97814420ae8292257deda5454fce3b"} Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.246211 4966 scope.go:117] "RemoveContainer" containerID="6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.285857 4966 scope.go:117] "RemoveContainer" containerID="c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.289353 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nrsr2"] Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.309923 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nrsr2"] Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.320091 4966 scope.go:117] "RemoveContainer" containerID="1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.351509 4966 scope.go:117] "RemoveContainer" containerID="6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb" Dec 17 09:07:28 crc kubenswrapper[4966]: E1217 09:07:28.352167 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb\": container with ID starting with 6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb not found: ID does not exist" containerID="6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.352218 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb"} err="failed to get container status \"6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb\": rpc error: code = NotFound desc = could not find container \"6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb\": container with ID starting with 6b123398d65a473fedec3bdbe3fe2316abc82b2bfd429889efa7c7f2622f7dfb not found: ID does not exist" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.352247 4966 scope.go:117] "RemoveContainer" containerID="c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa" Dec 17 09:07:28 crc kubenswrapper[4966]: E1217 09:07:28.353162 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa\": container with ID starting with c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa not found: ID does not exist" containerID="c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.353190 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa"} err="failed to get container status \"c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa\": rpc error: code = NotFound desc = could not find container \"c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa\": container with ID starting with c628181e1329fab5ebd4e91500d7d5c32299855b9077442a0ecda525e83aa5aa not found: ID does not exist" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.353211 4966 scope.go:117] "RemoveContainer" containerID="1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779" Dec 17 09:07:28 crc kubenswrapper[4966]: E1217 09:07:28.353447 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779\": container with ID starting with 1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779 not found: ID does not exist" containerID="1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.353474 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779"} err="failed to get container status \"1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779\": rpc error: code = NotFound desc = could not find container \"1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779\": container with ID starting with 1190b078e41d7872a27ba8ca24c7246271fd686fb7abe9908d53134c012b2779 not found: ID does not exist" Dec 17 09:07:28 crc kubenswrapper[4966]: I1217 09:07:28.841611 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" path="/var/lib/kubelet/pods/0a571f6a-6550-44e6-8cc7-17d01769d68a/volumes" Dec 17 09:07:40 crc kubenswrapper[4966]: I1217 09:07:40.346606 4966 generic.go:334] "Generic (PLEG): container finished" podID="c4bac84a-1f28-496b-91f2-c650fac7ed7d" containerID="723aeb1f2fa584a08b7547d4df3bda9b7920144df5a6b588b93ec444e43fd982" exitCode=0 Dec 17 09:07:40 crc kubenswrapper[4966]: I1217 09:07:40.347113 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" event={"ID":"c4bac84a-1f28-496b-91f2-c650fac7ed7d","Type":"ContainerDied","Data":"723aeb1f2fa584a08b7547d4df3bda9b7920144df5a6b588b93ec444e43fd982"} Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.063841 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.194980 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-ssh-key\") pod \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.195063 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-1\") pod \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.195095 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-1\") pod \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.195147 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-extra-config-0\") pod \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.195178 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-0\") pod \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.195206 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-combined-ca-bundle\") pod \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.195230 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-0\") pod \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.195375 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-inventory\") pod \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.195461 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hds4n\" (UniqueName: \"kubernetes.io/projected/c4bac84a-1f28-496b-91f2-c650fac7ed7d-kube-api-access-hds4n\") pod \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\" (UID: \"c4bac84a-1f28-496b-91f2-c650fac7ed7d\") " Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.203370 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4bac84a-1f28-496b-91f2-c650fac7ed7d-kube-api-access-hds4n" (OuterVolumeSpecName: "kube-api-access-hds4n") pod "c4bac84a-1f28-496b-91f2-c650fac7ed7d" (UID: "c4bac84a-1f28-496b-91f2-c650fac7ed7d"). InnerVolumeSpecName "kube-api-access-hds4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.205085 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c4bac84a-1f28-496b-91f2-c650fac7ed7d" (UID: "c4bac84a-1f28-496b-91f2-c650fac7ed7d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.253033 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "c4bac84a-1f28-496b-91f2-c650fac7ed7d" (UID: "c4bac84a-1f28-496b-91f2-c650fac7ed7d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.260092 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c4bac84a-1f28-496b-91f2-c650fac7ed7d" (UID: "c4bac84a-1f28-496b-91f2-c650fac7ed7d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.265684 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "c4bac84a-1f28-496b-91f2-c650fac7ed7d" (UID: "c4bac84a-1f28-496b-91f2-c650fac7ed7d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.298536 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hds4n\" (UniqueName: \"kubernetes.io/projected/c4bac84a-1f28-496b-91f2-c650fac7ed7d-kube-api-access-hds4n\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.298573 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.298584 4966 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.298597 4966 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.298610 4966 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.311050 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "c4bac84a-1f28-496b-91f2-c650fac7ed7d" (UID: "c4bac84a-1f28-496b-91f2-c650fac7ed7d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.312686 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "c4bac84a-1f28-496b-91f2-c650fac7ed7d" (UID: "c4bac84a-1f28-496b-91f2-c650fac7ed7d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.313102 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-inventory" (OuterVolumeSpecName: "inventory") pod "c4bac84a-1f28-496b-91f2-c650fac7ed7d" (UID: "c4bac84a-1f28-496b-91f2-c650fac7ed7d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.315023 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "c4bac84a-1f28-496b-91f2-c650fac7ed7d" (UID: "c4bac84a-1f28-496b-91f2-c650fac7ed7d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.364575 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" event={"ID":"c4bac84a-1f28-496b-91f2-c650fac7ed7d","Type":"ContainerDied","Data":"9f4f0eba91182c175927aa4c8e2ad64864c0668c848279d23ae979f62a5f032d"} Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.364617 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-gf4m7" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.364622 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f4f0eba91182c175927aa4c8e2ad64864c0668c848279d23ae979f62a5f032d" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.399992 4966 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.400022 4966 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.400032 4966 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.400055 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c4bac84a-1f28-496b-91f2-c650fac7ed7d-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.469214 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n"] Dec 17 09:07:42 crc kubenswrapper[4966]: E1217 09:07:42.469723 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="extract-utilities" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.469746 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="extract-utilities" Dec 17 09:07:42 crc kubenswrapper[4966]: E1217 09:07:42.469784 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="extract-content" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.469793 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="extract-content" Dec 17 09:07:42 crc kubenswrapper[4966]: E1217 09:07:42.469812 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="registry-server" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.469820 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="registry-server" Dec 17 09:07:42 crc kubenswrapper[4966]: E1217 09:07:42.469837 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bac84a-1f28-496b-91f2-c650fac7ed7d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.469846 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bac84a-1f28-496b-91f2-c650fac7ed7d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.470803 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a571f6a-6550-44e6-8cc7-17d01769d68a" containerName="registry-server" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.470832 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4bac84a-1f28-496b-91f2-c650fac7ed7d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.471619 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.477338 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.477565 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.477718 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.477978 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lb2xt" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.478228 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.487340 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n"] Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.501294 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.501334 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.501385 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.501401 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxzsj\" (UniqueName: \"kubernetes.io/projected/b221a164-40c6-41ee-84b8-1486d0b1bffb-kube-api-access-hxzsj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.501438 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.501498 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.501519 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.602723 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.603798 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.603989 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.604018 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.604112 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.604504 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxzsj\" (UniqueName: \"kubernetes.io/projected/b221a164-40c6-41ee-84b8-1486d0b1bffb-kube-api-access-hxzsj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.604581 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.623751 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.624156 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.624526 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.624929 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.628349 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.628448 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.635756 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxzsj\" (UniqueName: \"kubernetes.io/projected/b221a164-40c6-41ee-84b8-1486d0b1bffb-kube-api-access-hxzsj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:42 crc kubenswrapper[4966]: I1217 09:07:42.791345 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:07:43 crc kubenswrapper[4966]: I1217 09:07:43.363845 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n"] Dec 17 09:07:43 crc kubenswrapper[4966]: I1217 09:07:43.414448 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" event={"ID":"b221a164-40c6-41ee-84b8-1486d0b1bffb","Type":"ContainerStarted","Data":"be0951c726094db4ab6c9eaaf19622ae1b199a4d68f21371d89a2505da8640a8"} Dec 17 09:07:44 crc kubenswrapper[4966]: I1217 09:07:44.424533 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" event={"ID":"b221a164-40c6-41ee-84b8-1486d0b1bffb","Type":"ContainerStarted","Data":"4b7f4b8f726624cbe9c04e1b847cf5dddad118f5d3898a9167b8289e1a2395ac"} Dec 17 09:07:44 crc kubenswrapper[4966]: I1217 09:07:44.446127 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" podStartSLOduration=1.759304111 podStartE2EDuration="2.446109562s" podCreationTimestamp="2025-12-17 09:07:42 +0000 UTC" firstStartedPulling="2025-12-17 09:07:43.364520855 +0000 UTC m=+2798.909590797" lastFinishedPulling="2025-12-17 09:07:44.051326296 +0000 UTC m=+2799.596396248" observedRunningTime="2025-12-17 09:07:44.442292098 +0000 UTC m=+2799.987362040" watchObservedRunningTime="2025-12-17 09:07:44.446109562 +0000 UTC m=+2799.991179504" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.548167 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qsckv"] Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.551205 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.581995 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qsckv"] Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.681370 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-utilities\") pod \"redhat-marketplace-qsckv\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.681706 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dm5d\" (UniqueName: \"kubernetes.io/projected/b929f486-522c-491f-9be9-d291c04a18f3-kube-api-access-7dm5d\") pod \"redhat-marketplace-qsckv\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.681960 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-catalog-content\") pod \"redhat-marketplace-qsckv\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.785523 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dm5d\" (UniqueName: \"kubernetes.io/projected/b929f486-522c-491f-9be9-d291c04a18f3-kube-api-access-7dm5d\") pod \"redhat-marketplace-qsckv\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.785624 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-catalog-content\") pod \"redhat-marketplace-qsckv\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.785663 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-utilities\") pod \"redhat-marketplace-qsckv\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.786157 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-utilities\") pod \"redhat-marketplace-qsckv\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.786308 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-catalog-content\") pod \"redhat-marketplace-qsckv\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.831629 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dm5d\" (UniqueName: \"kubernetes.io/projected/b929f486-522c-491f-9be9-d291c04a18f3-kube-api-access-7dm5d\") pod \"redhat-marketplace-qsckv\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:13 crc kubenswrapper[4966]: I1217 09:08:13.875215 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:14 crc kubenswrapper[4966]: I1217 09:08:14.384854 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qsckv"] Dec 17 09:08:14 crc kubenswrapper[4966]: I1217 09:08:14.977219 4966 generic.go:334] "Generic (PLEG): container finished" podID="b929f486-522c-491f-9be9-d291c04a18f3" containerID="0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08" exitCode=0 Dec 17 09:08:14 crc kubenswrapper[4966]: I1217 09:08:14.977291 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qsckv" event={"ID":"b929f486-522c-491f-9be9-d291c04a18f3","Type":"ContainerDied","Data":"0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08"} Dec 17 09:08:14 crc kubenswrapper[4966]: I1217 09:08:14.977498 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qsckv" event={"ID":"b929f486-522c-491f-9be9-d291c04a18f3","Type":"ContainerStarted","Data":"51c500488f99d36a6a6a3f819edd140a428db67f21b9180304e4f6589bf07088"} Dec 17 09:08:16 crc kubenswrapper[4966]: E1217 09:08:16.804446 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb929f486_522c_491f_9be9_d291c04a18f3.slice/crio-6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb929f486_522c_491f_9be9_d291c04a18f3.slice/crio-conmon-6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628.scope\": RecentStats: unable to find data in memory cache]" Dec 17 09:08:17 crc kubenswrapper[4966]: I1217 09:08:17.029732 4966 generic.go:334] "Generic (PLEG): container finished" podID="b929f486-522c-491f-9be9-d291c04a18f3" containerID="6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628" exitCode=0 Dec 17 09:08:17 crc kubenswrapper[4966]: I1217 09:08:17.029818 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qsckv" event={"ID":"b929f486-522c-491f-9be9-d291c04a18f3","Type":"ContainerDied","Data":"6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628"} Dec 17 09:08:18 crc kubenswrapper[4966]: I1217 09:08:18.050045 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qsckv" event={"ID":"b929f486-522c-491f-9be9-d291c04a18f3","Type":"ContainerStarted","Data":"d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251"} Dec 17 09:08:18 crc kubenswrapper[4966]: I1217 09:08:18.077288 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qsckv" podStartSLOduration=2.41908877 podStartE2EDuration="5.077267877s" podCreationTimestamp="2025-12-17 09:08:13 +0000 UTC" firstStartedPulling="2025-12-17 09:08:14.978947117 +0000 UTC m=+2830.524017059" lastFinishedPulling="2025-12-17 09:08:17.637126164 +0000 UTC m=+2833.182196166" observedRunningTime="2025-12-17 09:08:18.071760387 +0000 UTC m=+2833.616830339" watchObservedRunningTime="2025-12-17 09:08:18.077267877 +0000 UTC m=+2833.622337829" Dec 17 09:08:23 crc kubenswrapper[4966]: I1217 09:08:23.876287 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:23 crc kubenswrapper[4966]: I1217 09:08:23.876916 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:23 crc kubenswrapper[4966]: I1217 09:08:23.923053 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:24 crc kubenswrapper[4966]: I1217 09:08:24.168231 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:24 crc kubenswrapper[4966]: I1217 09:08:24.221928 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qsckv"] Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.136282 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qsckv" podUID="b929f486-522c-491f-9be9-d291c04a18f3" containerName="registry-server" containerID="cri-o://d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251" gracePeriod=2 Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.618203 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.662755 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dm5d\" (UniqueName: \"kubernetes.io/projected/b929f486-522c-491f-9be9-d291c04a18f3-kube-api-access-7dm5d\") pod \"b929f486-522c-491f-9be9-d291c04a18f3\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.662919 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-utilities\") pod \"b929f486-522c-491f-9be9-d291c04a18f3\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.662973 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-catalog-content\") pod \"b929f486-522c-491f-9be9-d291c04a18f3\" (UID: \"b929f486-522c-491f-9be9-d291c04a18f3\") " Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.665037 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-utilities" (OuterVolumeSpecName: "utilities") pod "b929f486-522c-491f-9be9-d291c04a18f3" (UID: "b929f486-522c-491f-9be9-d291c04a18f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.671060 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b929f486-522c-491f-9be9-d291c04a18f3-kube-api-access-7dm5d" (OuterVolumeSpecName: "kube-api-access-7dm5d") pod "b929f486-522c-491f-9be9-d291c04a18f3" (UID: "b929f486-522c-491f-9be9-d291c04a18f3"). InnerVolumeSpecName "kube-api-access-7dm5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.694292 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b929f486-522c-491f-9be9-d291c04a18f3" (UID: "b929f486-522c-491f-9be9-d291c04a18f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.765682 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dm5d\" (UniqueName: \"kubernetes.io/projected/b929f486-522c-491f-9be9-d291c04a18f3-kube-api-access-7dm5d\") on node \"crc\" DevicePath \"\"" Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.765715 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:08:26 crc kubenswrapper[4966]: I1217 09:08:26.765726 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b929f486-522c-491f-9be9-d291c04a18f3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:08:27 crc kubenswrapper[4966]: E1217 09:08:27.054903 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb929f486_522c_491f_9be9_d291c04a18f3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb929f486_522c_491f_9be9_d291c04a18f3.slice/crio-51c500488f99d36a6a6a3f819edd140a428db67f21b9180304e4f6589bf07088\": RecentStats: unable to find data in memory cache]" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.146658 4966 generic.go:334] "Generic (PLEG): container finished" podID="b929f486-522c-491f-9be9-d291c04a18f3" containerID="d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251" exitCode=0 Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.146772 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qsckv" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.146803 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qsckv" event={"ID":"b929f486-522c-491f-9be9-d291c04a18f3","Type":"ContainerDied","Data":"d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251"} Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.147169 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qsckv" event={"ID":"b929f486-522c-491f-9be9-d291c04a18f3","Type":"ContainerDied","Data":"51c500488f99d36a6a6a3f819edd140a428db67f21b9180304e4f6589bf07088"} Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.147197 4966 scope.go:117] "RemoveContainer" containerID="d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.169349 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qsckv"] Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.183844 4966 scope.go:117] "RemoveContainer" containerID="6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.185978 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qsckv"] Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.206930 4966 scope.go:117] "RemoveContainer" containerID="0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.255526 4966 scope.go:117] "RemoveContainer" containerID="d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251" Dec 17 09:08:27 crc kubenswrapper[4966]: E1217 09:08:27.255980 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251\": container with ID starting with d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251 not found: ID does not exist" containerID="d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.256031 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251"} err="failed to get container status \"d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251\": rpc error: code = NotFound desc = could not find container \"d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251\": container with ID starting with d02ca042eeeb878d1d72b8e905e0d4cabc351bbdbfe09f4a3f926ece76050251 not found: ID does not exist" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.256059 4966 scope.go:117] "RemoveContainer" containerID="6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628" Dec 17 09:08:27 crc kubenswrapper[4966]: E1217 09:08:27.256686 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628\": container with ID starting with 6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628 not found: ID does not exist" containerID="6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.256708 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628"} err="failed to get container status \"6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628\": rpc error: code = NotFound desc = could not find container \"6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628\": container with ID starting with 6f137e01628cf67071c207f3ae054b97c87360c47d7a8ca1f312abef10418628 not found: ID does not exist" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.256722 4966 scope.go:117] "RemoveContainer" containerID="0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08" Dec 17 09:08:27 crc kubenswrapper[4966]: E1217 09:08:27.257069 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08\": container with ID starting with 0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08 not found: ID does not exist" containerID="0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08" Dec 17 09:08:27 crc kubenswrapper[4966]: I1217 09:08:27.257119 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08"} err="failed to get container status \"0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08\": rpc error: code = NotFound desc = could not find container \"0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08\": container with ID starting with 0bcc01c753fbe04b6817291fe0d5c36d0caddde4007dd43c14bc03b64ea7bf08 not found: ID does not exist" Dec 17 09:08:28 crc kubenswrapper[4966]: I1217 09:08:28.844347 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b929f486-522c-491f-9be9-d291c04a18f3" path="/var/lib/kubelet/pods/b929f486-522c-491f-9be9-d291c04a18f3/volumes" Dec 17 09:08:46 crc kubenswrapper[4966]: I1217 09:08:46.807956 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:08:46 crc kubenswrapper[4966]: I1217 09:08:46.808475 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:09:16 crc kubenswrapper[4966]: I1217 09:09:16.807902 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:09:16 crc kubenswrapper[4966]: I1217 09:09:16.808409 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:09:46 crc kubenswrapper[4966]: I1217 09:09:46.807617 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:09:46 crc kubenswrapper[4966]: I1217 09:09:46.808170 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:09:46 crc kubenswrapper[4966]: I1217 09:09:46.808216 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:09:46 crc kubenswrapper[4966]: I1217 09:09:46.808940 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"246ff596bfcf475a67b9d725efaf8f2795c418a5f24501a9f96ed1f28d380e19"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:09:46 crc kubenswrapper[4966]: I1217 09:09:46.808984 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://246ff596bfcf475a67b9d725efaf8f2795c418a5f24501a9f96ed1f28d380e19" gracePeriod=600 Dec 17 09:09:47 crc kubenswrapper[4966]: I1217 09:09:47.031174 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="246ff596bfcf475a67b9d725efaf8f2795c418a5f24501a9f96ed1f28d380e19" exitCode=0 Dec 17 09:09:47 crc kubenswrapper[4966]: I1217 09:09:47.031212 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"246ff596bfcf475a67b9d725efaf8f2795c418a5f24501a9f96ed1f28d380e19"} Dec 17 09:09:47 crc kubenswrapper[4966]: I1217 09:09:47.031240 4966 scope.go:117] "RemoveContainer" containerID="06ede3f4dd707981dcb62c84abed27158b85f1cd0f55fb64bd4e5d629a1d8592" Dec 17 09:09:48 crc kubenswrapper[4966]: I1217 09:09:48.042438 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52"} Dec 17 09:11:25 crc kubenswrapper[4966]: I1217 09:11:25.002564 4966 generic.go:334] "Generic (PLEG): container finished" podID="b221a164-40c6-41ee-84b8-1486d0b1bffb" containerID="4b7f4b8f726624cbe9c04e1b847cf5dddad118f5d3898a9167b8289e1a2395ac" exitCode=0 Dec 17 09:11:25 crc kubenswrapper[4966]: I1217 09:11:25.003038 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" event={"ID":"b221a164-40c6-41ee-84b8-1486d0b1bffb","Type":"ContainerDied","Data":"4b7f4b8f726624cbe9c04e1b847cf5dddad118f5d3898a9167b8289e1a2395ac"} Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.467635 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.560508 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-2\") pod \"b221a164-40c6-41ee-84b8-1486d0b1bffb\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.560584 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-1\") pod \"b221a164-40c6-41ee-84b8-1486d0b1bffb\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.560713 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-0\") pod \"b221a164-40c6-41ee-84b8-1486d0b1bffb\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.560754 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ssh-key\") pod \"b221a164-40c6-41ee-84b8-1486d0b1bffb\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.560800 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxzsj\" (UniqueName: \"kubernetes.io/projected/b221a164-40c6-41ee-84b8-1486d0b1bffb-kube-api-access-hxzsj\") pod \"b221a164-40c6-41ee-84b8-1486d0b1bffb\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.560990 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-inventory\") pod \"b221a164-40c6-41ee-84b8-1486d0b1bffb\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.561030 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-telemetry-combined-ca-bundle\") pod \"b221a164-40c6-41ee-84b8-1486d0b1bffb\" (UID: \"b221a164-40c6-41ee-84b8-1486d0b1bffb\") " Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.572347 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b221a164-40c6-41ee-84b8-1486d0b1bffb-kube-api-access-hxzsj" (OuterVolumeSpecName: "kube-api-access-hxzsj") pod "b221a164-40c6-41ee-84b8-1486d0b1bffb" (UID: "b221a164-40c6-41ee-84b8-1486d0b1bffb"). InnerVolumeSpecName "kube-api-access-hxzsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.573973 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b221a164-40c6-41ee-84b8-1486d0b1bffb" (UID: "b221a164-40c6-41ee-84b8-1486d0b1bffb"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.603769 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "b221a164-40c6-41ee-84b8-1486d0b1bffb" (UID: "b221a164-40c6-41ee-84b8-1486d0b1bffb"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.617318 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "b221a164-40c6-41ee-84b8-1486d0b1bffb" (UID: "b221a164-40c6-41ee-84b8-1486d0b1bffb"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.617483 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-inventory" (OuterVolumeSpecName: "inventory") pod "b221a164-40c6-41ee-84b8-1486d0b1bffb" (UID: "b221a164-40c6-41ee-84b8-1486d0b1bffb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.622294 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b221a164-40c6-41ee-84b8-1486d0b1bffb" (UID: "b221a164-40c6-41ee-84b8-1486d0b1bffb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.628957 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "b221a164-40c6-41ee-84b8-1486d0b1bffb" (UID: "b221a164-40c6-41ee-84b8-1486d0b1bffb"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.664302 4966 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.664346 4966 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.664359 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.664371 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxzsj\" (UniqueName: \"kubernetes.io/projected/b221a164-40c6-41ee-84b8-1486d0b1bffb-kube-api-access-hxzsj\") on node \"crc\" DevicePath \"\"" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.664419 4966 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-inventory\") on node \"crc\" DevicePath \"\"" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.664430 4966 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 09:11:26 crc kubenswrapper[4966]: I1217 09:11:26.664443 4966 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b221a164-40c6-41ee-84b8-1486d0b1bffb-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 17 09:11:27 crc kubenswrapper[4966]: I1217 09:11:27.026213 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" event={"ID":"b221a164-40c6-41ee-84b8-1486d0b1bffb","Type":"ContainerDied","Data":"be0951c726094db4ab6c9eaaf19622ae1b199a4d68f21371d89a2505da8640a8"} Dec 17 09:11:27 crc kubenswrapper[4966]: I1217 09:11:27.026541 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be0951c726094db4ab6c9eaaf19622ae1b199a4d68f21371d89a2505da8640a8" Dec 17 09:11:27 crc kubenswrapper[4966]: I1217 09:11:27.026589 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkb9n" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.706182 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cjfr7"] Dec 17 09:11:37 crc kubenswrapper[4966]: E1217 09:11:37.707066 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b929f486-522c-491f-9be9-d291c04a18f3" containerName="extract-utilities" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.707080 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b929f486-522c-491f-9be9-d291c04a18f3" containerName="extract-utilities" Dec 17 09:11:37 crc kubenswrapper[4966]: E1217 09:11:37.707104 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b929f486-522c-491f-9be9-d291c04a18f3" containerName="registry-server" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.707110 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b929f486-522c-491f-9be9-d291c04a18f3" containerName="registry-server" Dec 17 09:11:37 crc kubenswrapper[4966]: E1217 09:11:37.707122 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b221a164-40c6-41ee-84b8-1486d0b1bffb" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.707130 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b221a164-40c6-41ee-84b8-1486d0b1bffb" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 17 09:11:37 crc kubenswrapper[4966]: E1217 09:11:37.707148 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b929f486-522c-491f-9be9-d291c04a18f3" containerName="extract-content" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.707160 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="b929f486-522c-491f-9be9-d291c04a18f3" containerName="extract-content" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.707395 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b221a164-40c6-41ee-84b8-1486d0b1bffb" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.707419 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="b929f486-522c-491f-9be9-d291c04a18f3" containerName="registry-server" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.708888 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.723100 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjfr7"] Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.787539 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-utilities\") pod \"certified-operators-cjfr7\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.787600 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-catalog-content\") pod \"certified-operators-cjfr7\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.787662 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49sqh\" (UniqueName: \"kubernetes.io/projected/85c00dd1-6afb-4c05-996b-491d51c3063b-kube-api-access-49sqh\") pod \"certified-operators-cjfr7\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.889558 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-utilities\") pod \"certified-operators-cjfr7\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.889650 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-catalog-content\") pod \"certified-operators-cjfr7\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.889723 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49sqh\" (UniqueName: \"kubernetes.io/projected/85c00dd1-6afb-4c05-996b-491d51c3063b-kube-api-access-49sqh\") pod \"certified-operators-cjfr7\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.890177 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-utilities\") pod \"certified-operators-cjfr7\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.890185 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-catalog-content\") pod \"certified-operators-cjfr7\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:37 crc kubenswrapper[4966]: I1217 09:11:37.913941 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49sqh\" (UniqueName: \"kubernetes.io/projected/85c00dd1-6afb-4c05-996b-491d51c3063b-kube-api-access-49sqh\") pod \"certified-operators-cjfr7\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:38 crc kubenswrapper[4966]: I1217 09:11:38.081179 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:38 crc kubenswrapper[4966]: I1217 09:11:38.514228 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjfr7"] Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.141471 4966 generic.go:334] "Generic (PLEG): container finished" podID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerID="e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a" exitCode=0 Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.141546 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjfr7" event={"ID":"85c00dd1-6afb-4c05-996b-491d51c3063b","Type":"ContainerDied","Data":"e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a"} Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.141993 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjfr7" event={"ID":"85c00dd1-6afb-4c05-996b-491d51c3063b","Type":"ContainerStarted","Data":"2143d7934ce8a10819fcd816359e21e78b418fdf87640f82268cda73f669de6c"} Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.144138 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.502619 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ppk2z"] Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.509110 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.519240 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppk2z"] Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.641962 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrnfg\" (UniqueName: \"kubernetes.io/projected/67ff2290-8b2e-4ba6-90a6-45ac9318b528-kube-api-access-xrnfg\") pod \"community-operators-ppk2z\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.642156 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-catalog-content\") pod \"community-operators-ppk2z\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.642174 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-utilities\") pod \"community-operators-ppk2z\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.743987 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-catalog-content\") pod \"community-operators-ppk2z\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.744318 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-utilities\") pod \"community-operators-ppk2z\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.744378 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrnfg\" (UniqueName: \"kubernetes.io/projected/67ff2290-8b2e-4ba6-90a6-45ac9318b528-kube-api-access-xrnfg\") pod \"community-operators-ppk2z\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.744701 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-catalog-content\") pod \"community-operators-ppk2z\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.745075 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-utilities\") pod \"community-operators-ppk2z\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.767557 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrnfg\" (UniqueName: \"kubernetes.io/projected/67ff2290-8b2e-4ba6-90a6-45ac9318b528-kube-api-access-xrnfg\") pod \"community-operators-ppk2z\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:39 crc kubenswrapper[4966]: I1217 09:11:39.829117 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:40 crc kubenswrapper[4966]: I1217 09:11:40.532566 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ppk2z"] Dec 17 09:11:41 crc kubenswrapper[4966]: I1217 09:11:41.181346 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjfr7" event={"ID":"85c00dd1-6afb-4c05-996b-491d51c3063b","Type":"ContainerStarted","Data":"2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177"} Dec 17 09:11:41 crc kubenswrapper[4966]: I1217 09:11:41.182800 4966 generic.go:334] "Generic (PLEG): container finished" podID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerID="074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68" exitCode=0 Dec 17 09:11:41 crc kubenswrapper[4966]: I1217 09:11:41.182841 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppk2z" event={"ID":"67ff2290-8b2e-4ba6-90a6-45ac9318b528","Type":"ContainerDied","Data":"074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68"} Dec 17 09:11:41 crc kubenswrapper[4966]: I1217 09:11:41.182865 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppk2z" event={"ID":"67ff2290-8b2e-4ba6-90a6-45ac9318b528","Type":"ContainerStarted","Data":"e1c49d19343079364c3151bc0e2100ef9e11862d759003967d94e7897c31c88c"} Dec 17 09:11:42 crc kubenswrapper[4966]: I1217 09:11:42.195348 4966 generic.go:334] "Generic (PLEG): container finished" podID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerID="2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177" exitCode=0 Dec 17 09:11:42 crc kubenswrapper[4966]: I1217 09:11:42.195556 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjfr7" event={"ID":"85c00dd1-6afb-4c05-996b-491d51c3063b","Type":"ContainerDied","Data":"2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177"} Dec 17 09:11:43 crc kubenswrapper[4966]: I1217 09:11:43.222026 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppk2z" event={"ID":"67ff2290-8b2e-4ba6-90a6-45ac9318b528","Type":"ContainerStarted","Data":"588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205"} Dec 17 09:11:44 crc kubenswrapper[4966]: I1217 09:11:44.237915 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjfr7" event={"ID":"85c00dd1-6afb-4c05-996b-491d51c3063b","Type":"ContainerStarted","Data":"e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b"} Dec 17 09:11:44 crc kubenswrapper[4966]: I1217 09:11:44.241076 4966 generic.go:334] "Generic (PLEG): container finished" podID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerID="588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205" exitCode=0 Dec 17 09:11:44 crc kubenswrapper[4966]: I1217 09:11:44.241242 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppk2z" event={"ID":"67ff2290-8b2e-4ba6-90a6-45ac9318b528","Type":"ContainerDied","Data":"588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205"} Dec 17 09:11:44 crc kubenswrapper[4966]: I1217 09:11:44.285193 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cjfr7" podStartSLOduration=3.335724749 podStartE2EDuration="7.285177301s" podCreationTimestamp="2025-12-17 09:11:37 +0000 UTC" firstStartedPulling="2025-12-17 09:11:39.143880538 +0000 UTC m=+3034.688950480" lastFinishedPulling="2025-12-17 09:11:43.09333309 +0000 UTC m=+3038.638403032" observedRunningTime="2025-12-17 09:11:44.264168379 +0000 UTC m=+3039.809238311" watchObservedRunningTime="2025-12-17 09:11:44.285177301 +0000 UTC m=+3039.830247243" Dec 17 09:11:46 crc kubenswrapper[4966]: I1217 09:11:46.266459 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppk2z" event={"ID":"67ff2290-8b2e-4ba6-90a6-45ac9318b528","Type":"ContainerStarted","Data":"b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46"} Dec 17 09:11:46 crc kubenswrapper[4966]: I1217 09:11:46.284194 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ppk2z" podStartSLOduration=3.397878452 podStartE2EDuration="7.284177153s" podCreationTimestamp="2025-12-17 09:11:39 +0000 UTC" firstStartedPulling="2025-12-17 09:11:41.18397136 +0000 UTC m=+3036.729041332" lastFinishedPulling="2025-12-17 09:11:45.070270081 +0000 UTC m=+3040.615340033" observedRunningTime="2025-12-17 09:11:46.281281264 +0000 UTC m=+3041.826351206" watchObservedRunningTime="2025-12-17 09:11:46.284177153 +0000 UTC m=+3041.829247095" Dec 17 09:11:48 crc kubenswrapper[4966]: I1217 09:11:48.082364 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:48 crc kubenswrapper[4966]: I1217 09:11:48.082705 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:49 crc kubenswrapper[4966]: I1217 09:11:49.140345 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-cjfr7" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerName="registry-server" probeResult="failure" output=< Dec 17 09:11:49 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:11:49 crc kubenswrapper[4966]: > Dec 17 09:11:49 crc kubenswrapper[4966]: I1217 09:11:49.830020 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:49 crc kubenswrapper[4966]: I1217 09:11:49.830293 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:50 crc kubenswrapper[4966]: I1217 09:11:50.877376 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-ppk2z" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerName="registry-server" probeResult="failure" output=< Dec 17 09:11:50 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:11:50 crc kubenswrapper[4966]: > Dec 17 09:11:58 crc kubenswrapper[4966]: I1217 09:11:58.136714 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:58 crc kubenswrapper[4966]: I1217 09:11:58.192642 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:58 crc kubenswrapper[4966]: I1217 09:11:58.390967 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjfr7"] Dec 17 09:11:59 crc kubenswrapper[4966]: I1217 09:11:59.416956 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cjfr7" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerName="registry-server" containerID="cri-o://e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b" gracePeriod=2 Dec 17 09:11:59 crc kubenswrapper[4966]: I1217 09:11:59.915324 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:11:59 crc kubenswrapper[4966]: I1217 09:11:59.959699 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:11:59 crc kubenswrapper[4966]: I1217 09:11:59.997735 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.041471 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49sqh\" (UniqueName: \"kubernetes.io/projected/85c00dd1-6afb-4c05-996b-491d51c3063b-kube-api-access-49sqh\") pod \"85c00dd1-6afb-4c05-996b-491d51c3063b\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.041834 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-catalog-content\") pod \"85c00dd1-6afb-4c05-996b-491d51c3063b\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.042174 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-utilities\") pod \"85c00dd1-6afb-4c05-996b-491d51c3063b\" (UID: \"85c00dd1-6afb-4c05-996b-491d51c3063b\") " Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.042760 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-utilities" (OuterVolumeSpecName: "utilities") pod "85c00dd1-6afb-4c05-996b-491d51c3063b" (UID: "85c00dd1-6afb-4c05-996b-491d51c3063b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.044370 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.066797 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c00dd1-6afb-4c05-996b-491d51c3063b-kube-api-access-49sqh" (OuterVolumeSpecName: "kube-api-access-49sqh") pod "85c00dd1-6afb-4c05-996b-491d51c3063b" (UID: "85c00dd1-6afb-4c05-996b-491d51c3063b"). InnerVolumeSpecName "kube-api-access-49sqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.088629 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85c00dd1-6afb-4c05-996b-491d51c3063b" (UID: "85c00dd1-6afb-4c05-996b-491d51c3063b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.151205 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49sqh\" (UniqueName: \"kubernetes.io/projected/85c00dd1-6afb-4c05-996b-491d51c3063b-kube-api-access-49sqh\") on node \"crc\" DevicePath \"\"" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.151248 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c00dd1-6afb-4c05-996b-491d51c3063b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.432543 4966 generic.go:334] "Generic (PLEG): container finished" podID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerID="e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b" exitCode=0 Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.432630 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjfr7" event={"ID":"85c00dd1-6afb-4c05-996b-491d51c3063b","Type":"ContainerDied","Data":"e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b"} Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.434409 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjfr7" event={"ID":"85c00dd1-6afb-4c05-996b-491d51c3063b","Type":"ContainerDied","Data":"2143d7934ce8a10819fcd816359e21e78b418fdf87640f82268cda73f669de6c"} Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.432655 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjfr7" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.434496 4966 scope.go:117] "RemoveContainer" containerID="e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.522134 4966 scope.go:117] "RemoveContainer" containerID="2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.543931 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjfr7"] Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.582517 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cjfr7"] Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.613021 4966 scope.go:117] "RemoveContainer" containerID="e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.690944 4966 scope.go:117] "RemoveContainer" containerID="e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b" Dec 17 09:12:00 crc kubenswrapper[4966]: E1217 09:12:00.691444 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b\": container with ID starting with e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b not found: ID does not exist" containerID="e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.691507 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b"} err="failed to get container status \"e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b\": rpc error: code = NotFound desc = could not find container \"e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b\": container with ID starting with e3920927460623efe591bd97743e0584e9990b152d177c4d3f2264776428215b not found: ID does not exist" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.691534 4966 scope.go:117] "RemoveContainer" containerID="2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177" Dec 17 09:12:00 crc kubenswrapper[4966]: E1217 09:12:00.691814 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177\": container with ID starting with 2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177 not found: ID does not exist" containerID="2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.691865 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177"} err="failed to get container status \"2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177\": rpc error: code = NotFound desc = could not find container \"2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177\": container with ID starting with 2428040648e081c732f8017514d45034e0ff14eb0fbb73e832659224d4a01177 not found: ID does not exist" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.691906 4966 scope.go:117] "RemoveContainer" containerID="e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a" Dec 17 09:12:00 crc kubenswrapper[4966]: E1217 09:12:00.692233 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a\": container with ID starting with e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a not found: ID does not exist" containerID="e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.692264 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a"} err="failed to get container status \"e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a\": rpc error: code = NotFound desc = could not find container \"e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a\": container with ID starting with e9ef0f5aff3d4f6d9fd2856cf82d43c16e4df17532c232b4e464b0b47a8d785a not found: ID does not exist" Dec 17 09:12:00 crc kubenswrapper[4966]: I1217 09:12:00.841179 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" path="/var/lib/kubelet/pods/85c00dd1-6afb-4c05-996b-491d51c3063b/volumes" Dec 17 09:12:01 crc kubenswrapper[4966]: I1217 09:12:01.385148 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppk2z"] Dec 17 09:12:01 crc kubenswrapper[4966]: I1217 09:12:01.443816 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ppk2z" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerName="registry-server" containerID="cri-o://b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46" gracePeriod=2 Dec 17 09:12:01 crc kubenswrapper[4966]: I1217 09:12:01.987142 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.112266 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrnfg\" (UniqueName: \"kubernetes.io/projected/67ff2290-8b2e-4ba6-90a6-45ac9318b528-kube-api-access-xrnfg\") pod \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.112354 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-catalog-content\") pod \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.112466 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-utilities\") pod \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\" (UID: \"67ff2290-8b2e-4ba6-90a6-45ac9318b528\") " Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.113411 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-utilities" (OuterVolumeSpecName: "utilities") pod "67ff2290-8b2e-4ba6-90a6-45ac9318b528" (UID: "67ff2290-8b2e-4ba6-90a6-45ac9318b528"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.117225 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67ff2290-8b2e-4ba6-90a6-45ac9318b528-kube-api-access-xrnfg" (OuterVolumeSpecName: "kube-api-access-xrnfg") pod "67ff2290-8b2e-4ba6-90a6-45ac9318b528" (UID: "67ff2290-8b2e-4ba6-90a6-45ac9318b528"). InnerVolumeSpecName "kube-api-access-xrnfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.170221 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67ff2290-8b2e-4ba6-90a6-45ac9318b528" (UID: "67ff2290-8b2e-4ba6-90a6-45ac9318b528"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.215312 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrnfg\" (UniqueName: \"kubernetes.io/projected/67ff2290-8b2e-4ba6-90a6-45ac9318b528-kube-api-access-xrnfg\") on node \"crc\" DevicePath \"\"" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.215347 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.215362 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ff2290-8b2e-4ba6-90a6-45ac9318b528-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.455791 4966 generic.go:334] "Generic (PLEG): container finished" podID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerID="b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46" exitCode=0 Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.455841 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppk2z" event={"ID":"67ff2290-8b2e-4ba6-90a6-45ac9318b528","Type":"ContainerDied","Data":"b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46"} Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.455858 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ppk2z" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.455902 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ppk2z" event={"ID":"67ff2290-8b2e-4ba6-90a6-45ac9318b528","Type":"ContainerDied","Data":"e1c49d19343079364c3151bc0e2100ef9e11862d759003967d94e7897c31c88c"} Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.455925 4966 scope.go:117] "RemoveContainer" containerID="b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.486408 4966 scope.go:117] "RemoveContainer" containerID="588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.497951 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ppk2z"] Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.507673 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ppk2z"] Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.518887 4966 scope.go:117] "RemoveContainer" containerID="074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.559708 4966 scope.go:117] "RemoveContainer" containerID="b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46" Dec 17 09:12:02 crc kubenswrapper[4966]: E1217 09:12:02.560091 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46\": container with ID starting with b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46 not found: ID does not exist" containerID="b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.560120 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46"} err="failed to get container status \"b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46\": rpc error: code = NotFound desc = could not find container \"b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46\": container with ID starting with b9041b18d882c6f6e59523c8cde966be4411573759fb9ba519f7a49707237a46 not found: ID does not exist" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.560142 4966 scope.go:117] "RemoveContainer" containerID="588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205" Dec 17 09:12:02 crc kubenswrapper[4966]: E1217 09:12:02.560418 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205\": container with ID starting with 588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205 not found: ID does not exist" containerID="588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.560443 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205"} err="failed to get container status \"588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205\": rpc error: code = NotFound desc = could not find container \"588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205\": container with ID starting with 588e158eb67399caa8646b05f4f283aa544db7e990e53bb301cf7e115086b205 not found: ID does not exist" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.560460 4966 scope.go:117] "RemoveContainer" containerID="074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68" Dec 17 09:12:02 crc kubenswrapper[4966]: E1217 09:12:02.560733 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68\": container with ID starting with 074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68 not found: ID does not exist" containerID="074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.560761 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68"} err="failed to get container status \"074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68\": rpc error: code = NotFound desc = could not find container \"074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68\": container with ID starting with 074a54a194b9dde8b44afa9c6866140d0e5b7298240ce884d6ce300078ed4a68 not found: ID does not exist" Dec 17 09:12:02 crc kubenswrapper[4966]: I1217 09:12:02.848372 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" path="/var/lib/kubelet/pods/67ff2290-8b2e-4ba6-90a6-45ac9318b528/volumes" Dec 17 09:12:16 crc kubenswrapper[4966]: I1217 09:12:16.808063 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:12:16 crc kubenswrapper[4966]: I1217 09:12:16.808701 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.196612 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-multi-thread-testing"] Dec 17 09:12:17 crc kubenswrapper[4966]: E1217 09:12:17.197109 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerName="extract-utilities" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.197131 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerName="extract-utilities" Dec 17 09:12:17 crc kubenswrapper[4966]: E1217 09:12:17.197158 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerName="extract-utilities" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.197166 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerName="extract-utilities" Dec 17 09:12:17 crc kubenswrapper[4966]: E1217 09:12:17.197183 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerName="extract-content" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.197191 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerName="extract-content" Dec 17 09:12:17 crc kubenswrapper[4966]: E1217 09:12:17.197207 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerName="registry-server" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.197215 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerName="registry-server" Dec 17 09:12:17 crc kubenswrapper[4966]: E1217 09:12:17.197233 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerName="extract-content" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.197240 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerName="extract-content" Dec 17 09:12:17 crc kubenswrapper[4966]: E1217 09:12:17.197258 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerName="registry-server" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.197265 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerName="registry-server" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.197488 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ff2290-8b2e-4ba6-90a6-45ac9318b528" containerName="registry-server" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.197508 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c00dd1-6afb-4c05-996b-491d51c3063b" containerName="registry-server" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.198489 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.200133 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-j7vjx" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.203575 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.203921 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.206827 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.215071 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-multi-thread-testing"] Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.320468 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config-secret\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.320532 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-config-data\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.320581 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.320614 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2d4r\" (UniqueName: \"kubernetes.io/projected/780b74fd-a690-416e-80f1-3e28a75bd67c-kube-api-access-g2d4r\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.320646 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.320759 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.320829 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.321069 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ca-certs\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.321115 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ssh-key\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.422470 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.422600 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ca-certs\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.422645 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ssh-key\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.422681 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config-secret\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.422709 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-config-data\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.422743 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.422770 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2d4r\" (UniqueName: \"kubernetes.io/projected/780b74fd-a690-416e-80f1-3e28a75bd67c-kube-api-access-g2d4r\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.422798 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.422847 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.423013 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.424013 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-config-data\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.424253 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.424262 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.424998 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.429582 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ssh-key\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.433508 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config-secret\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.441615 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ca-certs\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.445668 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2d4r\" (UniqueName: \"kubernetes.io/projected/780b74fd-a690-416e-80f1-3e28a75bd67c-kube-api-access-g2d4r\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.460622 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:17 crc kubenswrapper[4966]: I1217 09:12:17.518982 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 09:12:18 crc kubenswrapper[4966]: I1217 09:12:18.119719 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-multi-thread-testing"] Dec 17 09:12:18 crc kubenswrapper[4966]: I1217 09:12:18.652759 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"780b74fd-a690-416e-80f1-3e28a75bd67c","Type":"ContainerStarted","Data":"a2cc6431d20ac839b5a4426d613ec41d80f5428efd6057cb2523ad49ccf32aae"} Dec 17 09:12:46 crc kubenswrapper[4966]: I1217 09:12:46.807353 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:12:46 crc kubenswrapper[4966]: I1217 09:12:46.807989 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:12:58 crc kubenswrapper[4966]: E1217 09:12:58.195195 4966 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-tempest-all:c3923531bcda0b0811b2d5053f189beb" Dec 17 09:12:58 crc kubenswrapper[4966]: E1217 09:12:58.195681 4966 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.156:5001/podified-antelope-centos9/openstack-tempest-all:c3923531bcda0b0811b2d5053f189beb" Dec 17 09:12:58 crc kubenswrapper[4966]: E1217 09:12:58.196897 4966 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:38.102.83.156:5001/podified-antelope-centos9/openstack-tempest-all:c3923531bcda0b0811b2d5053f189beb,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g2d4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-multi-thread-testing_openstack(780b74fd-a690-416e-80f1-3e28a75bd67c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 17 09:12:58 crc kubenswrapper[4966]: E1217 09:12:58.198072 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" podUID="780b74fd-a690-416e-80f1-3e28a75bd67c" Dec 17 09:12:59 crc kubenswrapper[4966]: E1217 09:12:59.101578 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.156:5001/podified-antelope-centos9/openstack-tempest-all:c3923531bcda0b0811b2d5053f189beb\\\"\"" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" podUID="780b74fd-a690-416e-80f1-3e28a75bd67c" Dec 17 09:13:12 crc kubenswrapper[4966]: I1217 09:13:12.966024 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 17 09:13:15 crc kubenswrapper[4966]: I1217 09:13:15.258300 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"780b74fd-a690-416e-80f1-3e28a75bd67c","Type":"ContainerStarted","Data":"0d516a9cf423ca68ecb9ba8e34eb110267cd67f48464edd92b0b1ce1a9e0a2f0"} Dec 17 09:13:15 crc kubenswrapper[4966]: I1217 09:13:15.278557 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" podStartSLOduration=4.452528631 podStartE2EDuration="59.278532755s" podCreationTimestamp="2025-12-17 09:12:16 +0000 UTC" firstStartedPulling="2025-12-17 09:12:18.135452042 +0000 UTC m=+3073.680521984" lastFinishedPulling="2025-12-17 09:13:12.961456126 +0000 UTC m=+3128.506526108" observedRunningTime="2025-12-17 09:13:15.27396745 +0000 UTC m=+3130.819037402" watchObservedRunningTime="2025-12-17 09:13:15.278532755 +0000 UTC m=+3130.823602707" Dec 17 09:13:16 crc kubenswrapper[4966]: I1217 09:13:16.807978 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:13:16 crc kubenswrapper[4966]: I1217 09:13:16.808258 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:13:16 crc kubenswrapper[4966]: I1217 09:13:16.808310 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:13:16 crc kubenswrapper[4966]: I1217 09:13:16.809038 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:13:16 crc kubenswrapper[4966]: I1217 09:13:16.809083 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" gracePeriod=600 Dec 17 09:13:16 crc kubenswrapper[4966]: E1217 09:13:16.929363 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:13:17 crc kubenswrapper[4966]: I1217 09:13:17.281639 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" exitCode=0 Dec 17 09:13:17 crc kubenswrapper[4966]: I1217 09:13:17.281719 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52"} Dec 17 09:13:17 crc kubenswrapper[4966]: I1217 09:13:17.281792 4966 scope.go:117] "RemoveContainer" containerID="246ff596bfcf475a67b9d725efaf8f2795c418a5f24501a9f96ed1f28d380e19" Dec 17 09:13:17 crc kubenswrapper[4966]: I1217 09:13:17.282775 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:13:17 crc kubenswrapper[4966]: E1217 09:13:17.283396 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:13:30 crc kubenswrapper[4966]: I1217 09:13:30.830284 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:13:30 crc kubenswrapper[4966]: E1217 09:13:30.831051 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:13:42 crc kubenswrapper[4966]: I1217 09:13:42.830656 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:13:42 crc kubenswrapper[4966]: E1217 09:13:42.831572 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:13:55 crc kubenswrapper[4966]: I1217 09:13:55.830718 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:13:55 crc kubenswrapper[4966]: E1217 09:13:55.831648 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:14:04 crc kubenswrapper[4966]: I1217 09:14:04.741282 4966 trace.go:236] Trace[1380871821]: "Calculate volume metrics of etcd-serving-ca for pod openshift-oauth-apiserver/apiserver-7bbb656c7d-vtwwx" (17-Dec-2025 09:14:03.663) (total time: 1056ms): Dec 17 09:14:04 crc kubenswrapper[4966]: Trace[1380871821]: [1.056069562s] [1.056069562s] END Dec 17 09:14:07 crc kubenswrapper[4966]: I1217 09:14:07.831507 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:14:07 crc kubenswrapper[4966]: E1217 09:14:07.832461 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:14:21 crc kubenswrapper[4966]: I1217 09:14:21.830984 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:14:21 crc kubenswrapper[4966]: E1217 09:14:21.831843 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:14:33 crc kubenswrapper[4966]: I1217 09:14:33.831263 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:14:33 crc kubenswrapper[4966]: E1217 09:14:33.832089 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:14:48 crc kubenswrapper[4966]: I1217 09:14:48.831695 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:14:48 crc kubenswrapper[4966]: E1217 09:14:48.832689 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.156687 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd"] Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.158804 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.162629 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.164327 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.171197 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd"] Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.194137 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxrrs\" (UniqueName: \"kubernetes.io/projected/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-kube-api-access-sxrrs\") pod \"collect-profiles-29432715-p66xd\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.194434 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-config-volume\") pod \"collect-profiles-29432715-p66xd\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.194534 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-secret-volume\") pod \"collect-profiles-29432715-p66xd\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.297194 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxrrs\" (UniqueName: \"kubernetes.io/projected/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-kube-api-access-sxrrs\") pod \"collect-profiles-29432715-p66xd\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.297261 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-config-volume\") pod \"collect-profiles-29432715-p66xd\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.297286 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-secret-volume\") pod \"collect-profiles-29432715-p66xd\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.298734 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-config-volume\") pod \"collect-profiles-29432715-p66xd\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.303212 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-secret-volume\") pod \"collect-profiles-29432715-p66xd\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.317144 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxrrs\" (UniqueName: \"kubernetes.io/projected/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-kube-api-access-sxrrs\") pod \"collect-profiles-29432715-p66xd\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:00 crc kubenswrapper[4966]: I1217 09:15:00.491252 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:02 crc kubenswrapper[4966]: I1217 09:15:02.205018 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd"] Dec 17 09:15:02 crc kubenswrapper[4966]: I1217 09:15:02.383962 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" event={"ID":"75b6bdc9-e860-4378-97cc-14c90d2b6e6f","Type":"ContainerStarted","Data":"acaf7e416d79aa4f7a14b70c0360a9db5e16ae29fc89e472f684ef3dd9faafac"} Dec 17 09:15:02 crc kubenswrapper[4966]: I1217 09:15:02.835661 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:15:02 crc kubenswrapper[4966]: E1217 09:15:02.836362 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:15:03 crc kubenswrapper[4966]: I1217 09:15:03.395051 4966 generic.go:334] "Generic (PLEG): container finished" podID="75b6bdc9-e860-4378-97cc-14c90d2b6e6f" containerID="dad663132e96f3e6f49d634e677d2bda89a09662e133194bf9b51c20c54be440" exitCode=0 Dec 17 09:15:03 crc kubenswrapper[4966]: I1217 09:15:03.395138 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" event={"ID":"75b6bdc9-e860-4378-97cc-14c90d2b6e6f","Type":"ContainerDied","Data":"dad663132e96f3e6f49d634e677d2bda89a09662e133194bf9b51c20c54be440"} Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.301294 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.413072 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-config-volume\") pod \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.413586 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxrrs\" (UniqueName: \"kubernetes.io/projected/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-kube-api-access-sxrrs\") pod \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.413806 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-secret-volume\") pod \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\" (UID: \"75b6bdc9-e860-4378-97cc-14c90d2b6e6f\") " Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.413587 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-config-volume" (OuterVolumeSpecName: "config-volume") pod "75b6bdc9-e860-4378-97cc-14c90d2b6e6f" (UID: "75b6bdc9-e860-4378-97cc-14c90d2b6e6f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.414546 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.418928 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" event={"ID":"75b6bdc9-e860-4378-97cc-14c90d2b6e6f","Type":"ContainerDied","Data":"acaf7e416d79aa4f7a14b70c0360a9db5e16ae29fc89e472f684ef3dd9faafac"} Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.418980 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acaf7e416d79aa4f7a14b70c0360a9db5e16ae29fc89e472f684ef3dd9faafac" Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.419619 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd" Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.420269 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "75b6bdc9-e860-4378-97cc-14c90d2b6e6f" (UID: "75b6bdc9-e860-4378-97cc-14c90d2b6e6f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.435539 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-kube-api-access-sxrrs" (OuterVolumeSpecName: "kube-api-access-sxrrs") pod "75b6bdc9-e860-4378-97cc-14c90d2b6e6f" (UID: "75b6bdc9-e860-4378-97cc-14c90d2b6e6f"). InnerVolumeSpecName "kube-api-access-sxrrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.516811 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 09:15:05 crc kubenswrapper[4966]: I1217 09:15:05.516854 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxrrs\" (UniqueName: \"kubernetes.io/projected/75b6bdc9-e860-4378-97cc-14c90d2b6e6f-kube-api-access-sxrrs\") on node \"crc\" DevicePath \"\"" Dec 17 09:15:06 crc kubenswrapper[4966]: I1217 09:15:06.402941 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw"] Dec 17 09:15:06 crc kubenswrapper[4966]: I1217 09:15:06.412629 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432670-wx7lw"] Dec 17 09:15:06 crc kubenswrapper[4966]: I1217 09:15:06.840942 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2" path="/var/lib/kubelet/pods/46fa2d36-361a-44c2-b3dc-e8dc0b6d80e2/volumes" Dec 17 09:15:13 crc kubenswrapper[4966]: I1217 09:15:13.831403 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:15:13 crc kubenswrapper[4966]: E1217 09:15:13.832217 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:15:26 crc kubenswrapper[4966]: I1217 09:15:26.830311 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:15:26 crc kubenswrapper[4966]: E1217 09:15:26.831146 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:15:38 crc kubenswrapper[4966]: I1217 09:15:38.831131 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:15:38 crc kubenswrapper[4966]: E1217 09:15:38.831962 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:15:52 crc kubenswrapper[4966]: I1217 09:15:52.830859 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:15:52 crc kubenswrapper[4966]: E1217 09:15:52.832727 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:15:58 crc kubenswrapper[4966]: I1217 09:15:58.198597 4966 scope.go:117] "RemoveContainer" containerID="90516996cc28c3c17310dc503dd6ff471ebabc6871623945ce7800777f33d954" Dec 17 09:16:03 crc kubenswrapper[4966]: I1217 09:16:03.830961 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:16:03 crc kubenswrapper[4966]: E1217 09:16:03.831743 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:16:16 crc kubenswrapper[4966]: I1217 09:16:16.903720 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:16:16 crc kubenswrapper[4966]: E1217 09:16:16.906685 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:16:29 crc kubenswrapper[4966]: I1217 09:16:29.830672 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:16:29 crc kubenswrapper[4966]: E1217 09:16:29.831577 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:16:43 crc kubenswrapper[4966]: I1217 09:16:43.830389 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:16:43 crc kubenswrapper[4966]: E1217 09:16:43.831171 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:16:56 crc kubenswrapper[4966]: I1217 09:16:56.831106 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:16:56 crc kubenswrapper[4966]: E1217 09:16:56.832066 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:17:09 crc kubenswrapper[4966]: I1217 09:17:09.830740 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:17:09 crc kubenswrapper[4966]: E1217 09:17:09.831512 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:17:21 crc kubenswrapper[4966]: I1217 09:17:21.830179 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:17:21 crc kubenswrapper[4966]: E1217 09:17:21.830982 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:17:22 crc kubenswrapper[4966]: I1217 09:17:22.248864 4966 trace.go:236] Trace[1201706716]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/redhat-marketplace-qxgxs" (17-Dec-2025 09:17:21.235) (total time: 1012ms): Dec 17 09:17:22 crc kubenswrapper[4966]: Trace[1201706716]: [1.012110195s] [1.012110195s] END Dec 17 09:17:35 crc kubenswrapper[4966]: I1217 09:17:35.831265 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:17:35 crc kubenswrapper[4966]: E1217 09:17:35.832173 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:17:50 crc kubenswrapper[4966]: I1217 09:17:50.831921 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:17:50 crc kubenswrapper[4966]: E1217 09:17:50.834054 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:18:02 crc kubenswrapper[4966]: I1217 09:18:02.830445 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:18:02 crc kubenswrapper[4966]: E1217 09:18:02.831268 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:18:07 crc kubenswrapper[4966]: I1217 09:18:07.844922 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-gvbzk" podUID="fcede706-7499-4c33-b80e-e1a13b7474dd" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.108006 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-brr62"] Dec 17 09:18:08 crc kubenswrapper[4966]: E1217 09:18:08.109909 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b6bdc9-e860-4378-97cc-14c90d2b6e6f" containerName="collect-profiles" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.109938 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b6bdc9-e860-4378-97cc-14c90d2b6e6f" containerName="collect-profiles" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.111705 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="75b6bdc9-e860-4378-97cc-14c90d2b6e6f" containerName="collect-profiles" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.114707 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.202797 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-utilities\") pod \"redhat-operators-brr62\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.202961 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9j2t\" (UniqueName: \"kubernetes.io/projected/a593fa3b-5fd7-4084-99bc-704d62cc2779-kube-api-access-c9j2t\") pod \"redhat-operators-brr62\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.203085 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-catalog-content\") pod \"redhat-operators-brr62\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.236444 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-brr62"] Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.310354 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-catalog-content\") pod \"redhat-operators-brr62\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.310630 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-utilities\") pod \"redhat-operators-brr62\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.310757 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9j2t\" (UniqueName: \"kubernetes.io/projected/a593fa3b-5fd7-4084-99bc-704d62cc2779-kube-api-access-c9j2t\") pod \"redhat-operators-brr62\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.317183 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-catalog-content\") pod \"redhat-operators-brr62\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.318731 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-utilities\") pod \"redhat-operators-brr62\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.375244 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9j2t\" (UniqueName: \"kubernetes.io/projected/a593fa3b-5fd7-4084-99bc-704d62cc2779-kube-api-access-c9j2t\") pod \"redhat-operators-brr62\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:08 crc kubenswrapper[4966]: I1217 09:18:08.439555 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:09 crc kubenswrapper[4966]: I1217 09:18:09.327642 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-brr62"] Dec 17 09:18:09 crc kubenswrapper[4966]: W1217 09:18:09.328185 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda593fa3b_5fd7_4084_99bc_704d62cc2779.slice/crio-965e9029cb859deeb845bcba58c853b29e06e8f68210096dd3541ec1a708ece2 WatchSource:0}: Error finding container 965e9029cb859deeb845bcba58c853b29e06e8f68210096dd3541ec1a708ece2: Status 404 returned error can't find the container with id 965e9029cb859deeb845bcba58c853b29e06e8f68210096dd3541ec1a708ece2 Dec 17 09:18:10 crc kubenswrapper[4966]: I1217 09:18:10.220237 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brr62" event={"ID":"a593fa3b-5fd7-4084-99bc-704d62cc2779","Type":"ContainerDied","Data":"2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8"} Dec 17 09:18:10 crc kubenswrapper[4966]: I1217 09:18:10.220269 4966 generic.go:334] "Generic (PLEG): container finished" podID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerID="2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8" exitCode=0 Dec 17 09:18:10 crc kubenswrapper[4966]: I1217 09:18:10.221112 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brr62" event={"ID":"a593fa3b-5fd7-4084-99bc-704d62cc2779","Type":"ContainerStarted","Data":"965e9029cb859deeb845bcba58c853b29e06e8f68210096dd3541ec1a708ece2"} Dec 17 09:18:10 crc kubenswrapper[4966]: I1217 09:18:10.231688 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 09:18:12 crc kubenswrapper[4966]: I1217 09:18:12.240345 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brr62" event={"ID":"a593fa3b-5fd7-4084-99bc-704d62cc2779","Type":"ContainerStarted","Data":"313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef"} Dec 17 09:18:14 crc kubenswrapper[4966]: I1217 09:18:14.977314 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:18:14 crc kubenswrapper[4966]: E1217 09:18:14.978032 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:18:16 crc kubenswrapper[4966]: I1217 09:18:16.287423 4966 generic.go:334] "Generic (PLEG): container finished" podID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerID="313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef" exitCode=0 Dec 17 09:18:16 crc kubenswrapper[4966]: I1217 09:18:16.287699 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brr62" event={"ID":"a593fa3b-5fd7-4084-99bc-704d62cc2779","Type":"ContainerDied","Data":"313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef"} Dec 17 09:18:17 crc kubenswrapper[4966]: I1217 09:18:17.299116 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brr62" event={"ID":"a593fa3b-5fd7-4084-99bc-704d62cc2779","Type":"ContainerStarted","Data":"98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b"} Dec 17 09:18:17 crc kubenswrapper[4966]: I1217 09:18:17.351543 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-brr62" podStartSLOduration=3.775577421 podStartE2EDuration="10.348213936s" podCreationTimestamp="2025-12-17 09:18:07 +0000 UTC" firstStartedPulling="2025-12-17 09:18:10.222402799 +0000 UTC m=+3425.767472731" lastFinishedPulling="2025-12-17 09:18:16.795039314 +0000 UTC m=+3432.340109246" observedRunningTime="2025-12-17 09:18:17.332262412 +0000 UTC m=+3432.877332364" watchObservedRunningTime="2025-12-17 09:18:17.348213936 +0000 UTC m=+3432.893283888" Dec 17 09:18:18 crc kubenswrapper[4966]: I1217 09:18:18.440157 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:18 crc kubenswrapper[4966]: I1217 09:18:18.440206 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:19 crc kubenswrapper[4966]: I1217 09:18:19.499420 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-brr62" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="registry-server" probeResult="failure" output=< Dec 17 09:18:19 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:18:19 crc kubenswrapper[4966]: > Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.451321 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s8c4v"] Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.454472 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.473404 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8c4v"] Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.518345 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjvjx\" (UniqueName: \"kubernetes.io/projected/e4f1a1d3-057c-419c-885a-4faa98e605b0-kube-api-access-wjvjx\") pod \"redhat-marketplace-s8c4v\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.519139 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-catalog-content\") pod \"redhat-marketplace-s8c4v\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.519221 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-utilities\") pod \"redhat-marketplace-s8c4v\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.621291 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjvjx\" (UniqueName: \"kubernetes.io/projected/e4f1a1d3-057c-419c-885a-4faa98e605b0-kube-api-access-wjvjx\") pod \"redhat-marketplace-s8c4v\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.621384 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-catalog-content\") pod \"redhat-marketplace-s8c4v\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.621425 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-utilities\") pod \"redhat-marketplace-s8c4v\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.621894 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-catalog-content\") pod \"redhat-marketplace-s8c4v\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.622101 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-utilities\") pod \"redhat-marketplace-s8c4v\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.647739 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjvjx\" (UniqueName: \"kubernetes.io/projected/e4f1a1d3-057c-419c-885a-4faa98e605b0-kube-api-access-wjvjx\") pod \"redhat-marketplace-s8c4v\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:20 crc kubenswrapper[4966]: I1217 09:18:20.808583 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:21 crc kubenswrapper[4966]: I1217 09:18:21.703624 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8c4v"] Dec 17 09:18:21 crc kubenswrapper[4966]: W1217 09:18:21.714695 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4f1a1d3_057c_419c_885a_4faa98e605b0.slice/crio-69993eefd6c89d5f5a5e84cfdcfab92ca3944a964266d74ee8e8d521d05b16b8 WatchSource:0}: Error finding container 69993eefd6c89d5f5a5e84cfdcfab92ca3944a964266d74ee8e8d521d05b16b8: Status 404 returned error can't find the container with id 69993eefd6c89d5f5a5e84cfdcfab92ca3944a964266d74ee8e8d521d05b16b8 Dec 17 09:18:22 crc kubenswrapper[4966]: I1217 09:18:22.356195 4966 generic.go:334] "Generic (PLEG): container finished" podID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerID="4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430" exitCode=0 Dec 17 09:18:22 crc kubenswrapper[4966]: I1217 09:18:22.356309 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8c4v" event={"ID":"e4f1a1d3-057c-419c-885a-4faa98e605b0","Type":"ContainerDied","Data":"4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430"} Dec 17 09:18:22 crc kubenswrapper[4966]: I1217 09:18:22.356542 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8c4v" event={"ID":"e4f1a1d3-057c-419c-885a-4faa98e605b0","Type":"ContainerStarted","Data":"69993eefd6c89d5f5a5e84cfdcfab92ca3944a964266d74ee8e8d521d05b16b8"} Dec 17 09:18:23 crc kubenswrapper[4966]: I1217 09:18:23.369823 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8c4v" event={"ID":"e4f1a1d3-057c-419c-885a-4faa98e605b0","Type":"ContainerStarted","Data":"a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359"} Dec 17 09:18:24 crc kubenswrapper[4966]: I1217 09:18:24.380106 4966 generic.go:334] "Generic (PLEG): container finished" podID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerID="a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359" exitCode=0 Dec 17 09:18:24 crc kubenswrapper[4966]: I1217 09:18:24.380161 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8c4v" event={"ID":"e4f1a1d3-057c-419c-885a-4faa98e605b0","Type":"ContainerDied","Data":"a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359"} Dec 17 09:18:25 crc kubenswrapper[4966]: I1217 09:18:25.391147 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8c4v" event={"ID":"e4f1a1d3-057c-419c-885a-4faa98e605b0","Type":"ContainerStarted","Data":"c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64"} Dec 17 09:18:25 crc kubenswrapper[4966]: I1217 09:18:25.414562 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s8c4v" podStartSLOduration=2.767219381 podStartE2EDuration="5.414542351s" podCreationTimestamp="2025-12-17 09:18:20 +0000 UTC" firstStartedPulling="2025-12-17 09:18:22.359216532 +0000 UTC m=+3437.904286474" lastFinishedPulling="2025-12-17 09:18:25.006539502 +0000 UTC m=+3440.551609444" observedRunningTime="2025-12-17 09:18:25.406582305 +0000 UTC m=+3440.951652257" watchObservedRunningTime="2025-12-17 09:18:25.414542351 +0000 UTC m=+3440.959612293" Dec 17 09:18:27 crc kubenswrapper[4966]: I1217 09:18:27.830829 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:18:28 crc kubenswrapper[4966]: I1217 09:18:28.456248 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"f593595a842f66f60712426dc4265d87e4da7c16085efe1318cfdaeb1f1a07f6"} Dec 17 09:18:29 crc kubenswrapper[4966]: I1217 09:18:29.513014 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-brr62" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="registry-server" probeResult="failure" output=< Dec 17 09:18:29 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:18:29 crc kubenswrapper[4966]: > Dec 17 09:18:30 crc kubenswrapper[4966]: I1217 09:18:30.809460 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:30 crc kubenswrapper[4966]: I1217 09:18:30.810391 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:31 crc kubenswrapper[4966]: I1217 09:18:31.873480 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-s8c4v" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerName="registry-server" probeResult="failure" output=< Dec 17 09:18:31 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:18:31 crc kubenswrapper[4966]: > Dec 17 09:18:39 crc kubenswrapper[4966]: I1217 09:18:39.494913 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-brr62" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="registry-server" probeResult="failure" output=< Dec 17 09:18:39 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:18:39 crc kubenswrapper[4966]: > Dec 17 09:18:40 crc kubenswrapper[4966]: I1217 09:18:40.883853 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:40 crc kubenswrapper[4966]: I1217 09:18:40.959541 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:41 crc kubenswrapper[4966]: I1217 09:18:41.139727 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8c4v"] Dec 17 09:18:42 crc kubenswrapper[4966]: I1217 09:18:42.587471 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s8c4v" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerName="registry-server" containerID="cri-o://c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64" gracePeriod=2 Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.466158 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.593783 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-catalog-content\") pod \"e4f1a1d3-057c-419c-885a-4faa98e605b0\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.593933 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-utilities\") pod \"e4f1a1d3-057c-419c-885a-4faa98e605b0\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.593998 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjvjx\" (UniqueName: \"kubernetes.io/projected/e4f1a1d3-057c-419c-885a-4faa98e605b0-kube-api-access-wjvjx\") pod \"e4f1a1d3-057c-419c-885a-4faa98e605b0\" (UID: \"e4f1a1d3-057c-419c-885a-4faa98e605b0\") " Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.598047 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-utilities" (OuterVolumeSpecName: "utilities") pod "e4f1a1d3-057c-419c-885a-4faa98e605b0" (UID: "e4f1a1d3-057c-419c-885a-4faa98e605b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.618006 4966 generic.go:334] "Generic (PLEG): container finished" podID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerID="c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64" exitCode=0 Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.618551 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8c4v" event={"ID":"e4f1a1d3-057c-419c-885a-4faa98e605b0","Type":"ContainerDied","Data":"c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64"} Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.618792 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s8c4v" event={"ID":"e4f1a1d3-057c-419c-885a-4faa98e605b0","Type":"ContainerDied","Data":"69993eefd6c89d5f5a5e84cfdcfab92ca3944a964266d74ee8e8d521d05b16b8"} Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.619212 4966 scope.go:117] "RemoveContainer" containerID="c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.619862 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s8c4v" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.627283 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f1a1d3-057c-419c-885a-4faa98e605b0-kube-api-access-wjvjx" (OuterVolumeSpecName: "kube-api-access-wjvjx") pod "e4f1a1d3-057c-419c-885a-4faa98e605b0" (UID: "e4f1a1d3-057c-419c-885a-4faa98e605b0"). InnerVolumeSpecName "kube-api-access-wjvjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.641958 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4f1a1d3-057c-419c-885a-4faa98e605b0" (UID: "e4f1a1d3-057c-419c-885a-4faa98e605b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.693889 4966 scope.go:117] "RemoveContainer" containerID="a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.702084 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.702207 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4f1a1d3-057c-419c-885a-4faa98e605b0-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.702227 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjvjx\" (UniqueName: \"kubernetes.io/projected/e4f1a1d3-057c-419c-885a-4faa98e605b0-kube-api-access-wjvjx\") on node \"crc\" DevicePath \"\"" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.732671 4966 scope.go:117] "RemoveContainer" containerID="4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.779565 4966 scope.go:117] "RemoveContainer" containerID="c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64" Dec 17 09:18:43 crc kubenswrapper[4966]: E1217 09:18:43.781948 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64\": container with ID starting with c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64 not found: ID does not exist" containerID="c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.781986 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64"} err="failed to get container status \"c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64\": rpc error: code = NotFound desc = could not find container \"c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64\": container with ID starting with c62361f49aac44b3a47a97e11c453e4ad3c34222bdd7b833560e010871f1bf64 not found: ID does not exist" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.782011 4966 scope.go:117] "RemoveContainer" containerID="a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359" Dec 17 09:18:43 crc kubenswrapper[4966]: E1217 09:18:43.782315 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359\": container with ID starting with a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359 not found: ID does not exist" containerID="a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.782333 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359"} err="failed to get container status \"a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359\": rpc error: code = NotFound desc = could not find container \"a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359\": container with ID starting with a08133e86f69d90602133e0bf551554df66b164587ab024d114c5025f30cf359 not found: ID does not exist" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.782358 4966 scope.go:117] "RemoveContainer" containerID="4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430" Dec 17 09:18:43 crc kubenswrapper[4966]: E1217 09:18:43.782599 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430\": container with ID starting with 4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430 not found: ID does not exist" containerID="4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.782618 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430"} err="failed to get container status \"4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430\": rpc error: code = NotFound desc = could not find container \"4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430\": container with ID starting with 4d97690497055ac7bd3f9526d50eb9de1decc1b9e7fe97fdf5cfad3fc6eb4430 not found: ID does not exist" Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.987245 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8c4v"] Dec 17 09:18:43 crc kubenswrapper[4966]: I1217 09:18:43.997277 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s8c4v"] Dec 17 09:18:44 crc kubenswrapper[4966]: I1217 09:18:44.844953 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" path="/var/lib/kubelet/pods/e4f1a1d3-057c-419c-885a-4faa98e605b0/volumes" Dec 17 09:18:48 crc kubenswrapper[4966]: I1217 09:18:48.504584 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:48 crc kubenswrapper[4966]: I1217 09:18:48.560979 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:48 crc kubenswrapper[4966]: I1217 09:18:48.764840 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-brr62"] Dec 17 09:18:49 crc kubenswrapper[4966]: I1217 09:18:49.700702 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-brr62" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="registry-server" containerID="cri-o://98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b" gracePeriod=2 Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.388517 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.544680 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9j2t\" (UniqueName: \"kubernetes.io/projected/a593fa3b-5fd7-4084-99bc-704d62cc2779-kube-api-access-c9j2t\") pod \"a593fa3b-5fd7-4084-99bc-704d62cc2779\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.544844 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-utilities\") pod \"a593fa3b-5fd7-4084-99bc-704d62cc2779\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.544938 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-catalog-content\") pod \"a593fa3b-5fd7-4084-99bc-704d62cc2779\" (UID: \"a593fa3b-5fd7-4084-99bc-704d62cc2779\") " Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.546414 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-utilities" (OuterVolumeSpecName: "utilities") pod "a593fa3b-5fd7-4084-99bc-704d62cc2779" (UID: "a593fa3b-5fd7-4084-99bc-704d62cc2779"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.554199 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a593fa3b-5fd7-4084-99bc-704d62cc2779-kube-api-access-c9j2t" (OuterVolumeSpecName: "kube-api-access-c9j2t") pod "a593fa3b-5fd7-4084-99bc-704d62cc2779" (UID: "a593fa3b-5fd7-4084-99bc-704d62cc2779"). InnerVolumeSpecName "kube-api-access-c9j2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.647813 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.647846 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9j2t\" (UniqueName: \"kubernetes.io/projected/a593fa3b-5fd7-4084-99bc-704d62cc2779-kube-api-access-c9j2t\") on node \"crc\" DevicePath \"\"" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.651315 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a593fa3b-5fd7-4084-99bc-704d62cc2779" (UID: "a593fa3b-5fd7-4084-99bc-704d62cc2779"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.710788 4966 generic.go:334] "Generic (PLEG): container finished" podID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerID="98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b" exitCode=0 Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.710837 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brr62" event={"ID":"a593fa3b-5fd7-4084-99bc-704d62cc2779","Type":"ContainerDied","Data":"98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b"} Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.710875 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brr62" event={"ID":"a593fa3b-5fd7-4084-99bc-704d62cc2779","Type":"ContainerDied","Data":"965e9029cb859deeb845bcba58c853b29e06e8f68210096dd3541ec1a708ece2"} Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.710906 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brr62" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.710922 4966 scope.go:117] "RemoveContainer" containerID="98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.751070 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a593fa3b-5fd7-4084-99bc-704d62cc2779-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.753250 4966 scope.go:117] "RemoveContainer" containerID="313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.756800 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-brr62"] Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.769448 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-brr62"] Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.783460 4966 scope.go:117] "RemoveContainer" containerID="2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.833104 4966 scope.go:117] "RemoveContainer" containerID="98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b" Dec 17 09:18:50 crc kubenswrapper[4966]: E1217 09:18:50.834421 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b\": container with ID starting with 98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b not found: ID does not exist" containerID="98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.834453 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b"} err="failed to get container status \"98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b\": rpc error: code = NotFound desc = could not find container \"98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b\": container with ID starting with 98846291079a715a3e977b9e626f89cd18ea37c32a5f0c7a0826bec7f9158a1b not found: ID does not exist" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.834473 4966 scope.go:117] "RemoveContainer" containerID="313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef" Dec 17 09:18:50 crc kubenswrapper[4966]: E1217 09:18:50.834923 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef\": container with ID starting with 313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef not found: ID does not exist" containerID="313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.834945 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef"} err="failed to get container status \"313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef\": rpc error: code = NotFound desc = could not find container \"313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef\": container with ID starting with 313b83f8a3eca0608ea6bc1378abae1845dee72931420d0d56ccd4c88ebcfaef not found: ID does not exist" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.834958 4966 scope.go:117] "RemoveContainer" containerID="2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8" Dec 17 09:18:50 crc kubenswrapper[4966]: E1217 09:18:50.835285 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8\": container with ID starting with 2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8 not found: ID does not exist" containerID="2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.835309 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8"} err="failed to get container status \"2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8\": rpc error: code = NotFound desc = could not find container \"2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8\": container with ID starting with 2061d42cf491608912e322c99d0b8cc05b69cfc047f95cda75dffcc0ed2788c8 not found: ID does not exist" Dec 17 09:18:50 crc kubenswrapper[4966]: I1217 09:18:50.842327 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" path="/var/lib/kubelet/pods/a593fa3b-5fd7-4084-99bc-704d62cc2779/volumes" Dec 17 09:20:37 crc kubenswrapper[4966]: I1217 09:20:37.556929 4966 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.167759645s: [/var/lib/containers/storage/overlay/98f017997772a02bdd730c4df3cd8b50f2207965f7eeba8febbe66b022632c09/diff /var/log/pods/openshift-network-console_networking-console-plugin-85b44fc459-gdk6g_5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/networking-console-plugin/0.log]; will not log again for this container unless duration exceeds 2s Dec 17 09:20:37 crc kubenswrapper[4966]: I1217 09:20:37.599480 4966 trace.go:236] Trace[1734804404]: "Calculate volume metrics of signing-cabundle for pod openshift-service-ca/service-ca-9c57cc56f-8j46g" (17-Dec-2025 09:20:35.703) (total time: 1895ms): Dec 17 09:20:37 crc kubenswrapper[4966]: Trace[1734804404]: [1.895186591s] [1.895186591s] END Dec 17 09:20:37 crc kubenswrapper[4966]: I1217 09:20:37.599525 4966 trace.go:236] Trace[197110031]: "Calculate volume metrics of etcd-service-ca for pod openshift-etcd-operator/etcd-operator-b45778765-9pzvc" (17-Dec-2025 09:20:36.452) (total time: 1136ms): Dec 17 09:20:37 crc kubenswrapper[4966]: Trace[197110031]: [1.136754381s] [1.136754381s] END Dec 17 09:20:37 crc kubenswrapper[4966]: I1217 09:20:37.600748 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7b67c7f6c5x6jrd" podUID="add4eb97-35a5-4070-ac9f-e14eb5741aac" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.87:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 09:20:37 crc kubenswrapper[4966]: I1217 09:20:37.644134 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-gvbzk" podUID="fcede706-7499-4c33-b80e-e1a13b7474dd" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 09:20:38 crc kubenswrapper[4966]: I1217 09:20:38.044946 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" podUID="46dfa510-96af-4d45-9be7-5bd9a8588a61" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 09:20:46 crc kubenswrapper[4966]: I1217 09:20:46.808073 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:20:46 crc kubenswrapper[4966]: I1217 09:20:46.808729 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:21:15 crc kubenswrapper[4966]: I1217 09:21:15.766707 4966 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.06507462s: [/var/lib/containers/storage/overlay/6dbd78f49e3a9ea65edb046282cdfc58535ba513a6be283ba5635209fb4c5d11/diff /var/log/pods/openstack_barbican-worker-564669fcf5-gv2b4_f5a721a1-6544-4175-9734-232023ca3f2f/barbican-worker/0.log]; will not log again for this container unless duration exceeds 2s Dec 17 09:21:15 crc kubenswrapper[4966]: I1217 09:21:15.769241 4966 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.21125434s: [/var/lib/containers/storage/overlay/4df9b20a1f24ef92aa6ae0dc9ccf1e1700862282d230519f0c31dc558f3bd6d7/diff /var/log/pods/openstack_barbican-api-589dbc6cbd-4hskt_b794efa0-cd44-46ef-bd39-5bcde98ec271/barbican-api/0.log]; will not log again for this container unless duration exceeds 2s Dec 17 09:21:16 crc kubenswrapper[4966]: I1217 09:21:16.807206 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:21:16 crc kubenswrapper[4966]: I1217 09:21:16.807253 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.891798 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pmssb"] Dec 17 09:21:37 crc kubenswrapper[4966]: E1217 09:21:37.893650 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="registry-server" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.893684 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="registry-server" Dec 17 09:21:37 crc kubenswrapper[4966]: E1217 09:21:37.893797 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="extract-content" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.893844 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="extract-content" Dec 17 09:21:37 crc kubenswrapper[4966]: E1217 09:21:37.893852 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerName="extract-content" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.893859 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerName="extract-content" Dec 17 09:21:37 crc kubenswrapper[4966]: E1217 09:21:37.893865 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="extract-utilities" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.893897 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="extract-utilities" Dec 17 09:21:37 crc kubenswrapper[4966]: E1217 09:21:37.893914 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerName="extract-utilities" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.893919 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerName="extract-utilities" Dec 17 09:21:37 crc kubenswrapper[4966]: E1217 09:21:37.893939 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerName="registry-server" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.893948 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerName="registry-server" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.894390 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a593fa3b-5fd7-4084-99bc-704d62cc2779" containerName="registry-server" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.894413 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f1a1d3-057c-419c-885a-4faa98e605b0" containerName="registry-server" Dec 17 09:21:37 crc kubenswrapper[4966]: I1217 09:21:37.906699 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.050180 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-utilities\") pod \"certified-operators-pmssb\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.050549 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf7xw\" (UniqueName: \"kubernetes.io/projected/63fee3e3-d38a-445c-83c5-91bde266b4b5-kube-api-access-qf7xw\") pod \"certified-operators-pmssb\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.050937 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-catalog-content\") pod \"certified-operators-pmssb\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.137633 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pmssb"] Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.152825 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-utilities\") pod \"certified-operators-pmssb\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.152945 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf7xw\" (UniqueName: \"kubernetes.io/projected/63fee3e3-d38a-445c-83c5-91bde266b4b5-kube-api-access-qf7xw\") pod \"certified-operators-pmssb\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.153007 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-catalog-content\") pod \"certified-operators-pmssb\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.154978 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-catalog-content\") pod \"certified-operators-pmssb\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.156500 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-utilities\") pod \"certified-operators-pmssb\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.204855 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf7xw\" (UniqueName: \"kubernetes.io/projected/63fee3e3-d38a-445c-83c5-91bde266b4b5-kube-api-access-qf7xw\") pod \"certified-operators-pmssb\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:38 crc kubenswrapper[4966]: I1217 09:21:38.364149 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:39 crc kubenswrapper[4966]: I1217 09:21:39.462780 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pmssb"] Dec 17 09:21:39 crc kubenswrapper[4966]: W1217 09:21:39.473713 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63fee3e3_d38a_445c_83c5_91bde266b4b5.slice/crio-ef4a6f692db047df544fc830a7cdf7b2434fcfd35174939831ddfd9ab4e8a83b WatchSource:0}: Error finding container ef4a6f692db047df544fc830a7cdf7b2434fcfd35174939831ddfd9ab4e8a83b: Status 404 returned error can't find the container with id ef4a6f692db047df544fc830a7cdf7b2434fcfd35174939831ddfd9ab4e8a83b Dec 17 09:21:40 crc kubenswrapper[4966]: I1217 09:21:40.048528 4966 generic.go:334] "Generic (PLEG): container finished" podID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerID="878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72" exitCode=0 Dec 17 09:21:40 crc kubenswrapper[4966]: I1217 09:21:40.048649 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmssb" event={"ID":"63fee3e3-d38a-445c-83c5-91bde266b4b5","Type":"ContainerDied","Data":"878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72"} Dec 17 09:21:40 crc kubenswrapper[4966]: I1217 09:21:40.048848 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmssb" event={"ID":"63fee3e3-d38a-445c-83c5-91bde266b4b5","Type":"ContainerStarted","Data":"ef4a6f692db047df544fc830a7cdf7b2434fcfd35174939831ddfd9ab4e8a83b"} Dec 17 09:21:42 crc kubenswrapper[4966]: I1217 09:21:42.068003 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmssb" event={"ID":"63fee3e3-d38a-445c-83c5-91bde266b4b5","Type":"ContainerStarted","Data":"e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4"} Dec 17 09:21:44 crc kubenswrapper[4966]: I1217 09:21:44.096097 4966 generic.go:334] "Generic (PLEG): container finished" podID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerID="e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4" exitCode=0 Dec 17 09:21:44 crc kubenswrapper[4966]: I1217 09:21:44.096156 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmssb" event={"ID":"63fee3e3-d38a-445c-83c5-91bde266b4b5","Type":"ContainerDied","Data":"e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4"} Dec 17 09:21:45 crc kubenswrapper[4966]: I1217 09:21:45.108929 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmssb" event={"ID":"63fee3e3-d38a-445c-83c5-91bde266b4b5","Type":"ContainerStarted","Data":"af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345"} Dec 17 09:21:45 crc kubenswrapper[4966]: I1217 09:21:45.139326 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pmssb" podStartSLOduration=3.596943821 podStartE2EDuration="8.135027461s" podCreationTimestamp="2025-12-17 09:21:37 +0000 UTC" firstStartedPulling="2025-12-17 09:21:40.050444701 +0000 UTC m=+3635.595514643" lastFinishedPulling="2025-12-17 09:21:44.588528341 +0000 UTC m=+3640.133598283" observedRunningTime="2025-12-17 09:21:45.130675112 +0000 UTC m=+3640.675745054" watchObservedRunningTime="2025-12-17 09:21:45.135027461 +0000 UTC m=+3640.680097403" Dec 17 09:21:46 crc kubenswrapper[4966]: I1217 09:21:46.807846 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:21:46 crc kubenswrapper[4966]: I1217 09:21:46.808223 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:21:46 crc kubenswrapper[4966]: I1217 09:21:46.808266 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:21:46 crc kubenswrapper[4966]: I1217 09:21:46.809243 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f593595a842f66f60712426dc4265d87e4da7c16085efe1318cfdaeb1f1a07f6"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:21:46 crc kubenswrapper[4966]: I1217 09:21:46.809301 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://f593595a842f66f60712426dc4265d87e4da7c16085efe1318cfdaeb1f1a07f6" gracePeriod=600 Dec 17 09:21:47 crc kubenswrapper[4966]: I1217 09:21:47.128509 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="f593595a842f66f60712426dc4265d87e4da7c16085efe1318cfdaeb1f1a07f6" exitCode=0 Dec 17 09:21:47 crc kubenswrapper[4966]: I1217 09:21:47.128584 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"f593595a842f66f60712426dc4265d87e4da7c16085efe1318cfdaeb1f1a07f6"} Dec 17 09:21:47 crc kubenswrapper[4966]: I1217 09:21:47.129428 4966 scope.go:117] "RemoveContainer" containerID="698087a504dd01a74b10179ad11eadadb9e55280572e49ff2baecbe378c5ee52" Dec 17 09:21:48 crc kubenswrapper[4966]: I1217 09:21:48.141799 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e"} Dec 17 09:21:48 crc kubenswrapper[4966]: I1217 09:21:48.364571 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:48 crc kubenswrapper[4966]: I1217 09:21:48.364624 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:49 crc kubenswrapper[4966]: I1217 09:21:49.433613 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-pmssb" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerName="registry-server" probeResult="failure" output=< Dec 17 09:21:49 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:21:49 crc kubenswrapper[4966]: > Dec 17 09:21:58 crc kubenswrapper[4966]: I1217 09:21:58.423534 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:58 crc kubenswrapper[4966]: I1217 09:21:58.487618 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:21:58 crc kubenswrapper[4966]: I1217 09:21:58.731181 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pmssb"] Dec 17 09:22:00 crc kubenswrapper[4966]: I1217 09:22:00.244003 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pmssb" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerName="registry-server" containerID="cri-o://af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345" gracePeriod=2 Dec 17 09:22:00 crc kubenswrapper[4966]: E1217 09:22:00.435919 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63fee3e3_d38a_445c_83c5_91bde266b4b5.slice/crio-conmon-af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63fee3e3_d38a_445c_83c5_91bde266b4b5.slice/crio-af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345.scope\": RecentStats: unable to find data in memory cache]" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.237402 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.252444 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmssb" event={"ID":"63fee3e3-d38a-445c-83c5-91bde266b4b5","Type":"ContainerDied","Data":"af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345"} Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.252380 4966 generic.go:334] "Generic (PLEG): container finished" podID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerID="af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345" exitCode=0 Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.252626 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pmssb" event={"ID":"63fee3e3-d38a-445c-83c5-91bde266b4b5","Type":"ContainerDied","Data":"ef4a6f692db047df544fc830a7cdf7b2434fcfd35174939831ddfd9ab4e8a83b"} Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.252597 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pmssb" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.252758 4966 scope.go:117] "RemoveContainer" containerID="af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.296832 4966 scope.go:117] "RemoveContainer" containerID="e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.320944 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf7xw\" (UniqueName: \"kubernetes.io/projected/63fee3e3-d38a-445c-83c5-91bde266b4b5-kube-api-access-qf7xw\") pod \"63fee3e3-d38a-445c-83c5-91bde266b4b5\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.321012 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-utilities\") pod \"63fee3e3-d38a-445c-83c5-91bde266b4b5\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.320947 4966 scope.go:117] "RemoveContainer" containerID="878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.321140 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-catalog-content\") pod \"63fee3e3-d38a-445c-83c5-91bde266b4b5\" (UID: \"63fee3e3-d38a-445c-83c5-91bde266b4b5\") " Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.323571 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-utilities" (OuterVolumeSpecName: "utilities") pod "63fee3e3-d38a-445c-83c5-91bde266b4b5" (UID: "63fee3e3-d38a-445c-83c5-91bde266b4b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.348120 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63fee3e3-d38a-445c-83c5-91bde266b4b5-kube-api-access-qf7xw" (OuterVolumeSpecName: "kube-api-access-qf7xw") pod "63fee3e3-d38a-445c-83c5-91bde266b4b5" (UID: "63fee3e3-d38a-445c-83c5-91bde266b4b5"). InnerVolumeSpecName "kube-api-access-qf7xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.390733 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63fee3e3-d38a-445c-83c5-91bde266b4b5" (UID: "63fee3e3-d38a-445c-83c5-91bde266b4b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.408259 4966 scope.go:117] "RemoveContainer" containerID="af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345" Dec 17 09:22:01 crc kubenswrapper[4966]: E1217 09:22:01.410051 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345\": container with ID starting with af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345 not found: ID does not exist" containerID="af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.410292 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345"} err="failed to get container status \"af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345\": rpc error: code = NotFound desc = could not find container \"af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345\": container with ID starting with af336cbb99cdd0bd847108bcd776695dfcaae8485b52affd1b7d4964179ee345 not found: ID does not exist" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.410319 4966 scope.go:117] "RemoveContainer" containerID="e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4" Dec 17 09:22:01 crc kubenswrapper[4966]: E1217 09:22:01.411170 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4\": container with ID starting with e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4 not found: ID does not exist" containerID="e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.411283 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4"} err="failed to get container status \"e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4\": rpc error: code = NotFound desc = could not find container \"e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4\": container with ID starting with e5e6f34fd53e41e39ac06e55b665d2e11311f0a1253827740a4d63b05499ecf4 not found: ID does not exist" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.411412 4966 scope.go:117] "RemoveContainer" containerID="878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72" Dec 17 09:22:01 crc kubenswrapper[4966]: E1217 09:22:01.411690 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72\": container with ID starting with 878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72 not found: ID does not exist" containerID="878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.411711 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72"} err="failed to get container status \"878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72\": rpc error: code = NotFound desc = could not find container \"878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72\": container with ID starting with 878740fa08d5d00d200ec1815c80fe62ccc570dff5b4ad2cf600bc1af6d17e72 not found: ID does not exist" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.423007 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf7xw\" (UniqueName: \"kubernetes.io/projected/63fee3e3-d38a-445c-83c5-91bde266b4b5-kube-api-access-qf7xw\") on node \"crc\" DevicePath \"\"" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.423210 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.423293 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63fee3e3-d38a-445c-83c5-91bde266b4b5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.598995 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pmssb"] Dec 17 09:22:01 crc kubenswrapper[4966]: I1217 09:22:01.612290 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pmssb"] Dec 17 09:22:02 crc kubenswrapper[4966]: I1217 09:22:02.847184 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" path="/var/lib/kubelet/pods/63fee3e3-d38a-445c-83c5-91bde266b4b5/volumes" Dec 17 09:22:17 crc kubenswrapper[4966]: I1217 09:22:17.388846 4966 trace.go:236] Trace[1335372291]: "Calculate volume metrics of metallb-excludel2 for pod metallb-system/speaker-8mc27" (17-Dec-2025 09:22:16.299) (total time: 1087ms): Dec 17 09:22:17 crc kubenswrapper[4966]: Trace[1335372291]: [1.087501939s] [1.087501939s] END Dec 17 09:22:17 crc kubenswrapper[4966]: I1217 09:22:17.400772 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-gvbzk" podUID="fcede706-7499-4c33-b80e-e1a13b7474dd" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 09:24:16 crc kubenswrapper[4966]: I1217 09:24:16.810724 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:24:16 crc kubenswrapper[4966]: I1217 09:24:16.814416 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:24:46 crc kubenswrapper[4966]: I1217 09:24:46.808103 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:24:46 crc kubenswrapper[4966]: I1217 09:24:46.808610 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:25:16 crc kubenswrapper[4966]: I1217 09:25:16.807790 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:25:17 crc kubenswrapper[4966]: I1217 09:25:16.808355 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:25:17 crc kubenswrapper[4966]: I1217 09:25:17.132321 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:25:17 crc kubenswrapper[4966]: I1217 09:25:17.149115 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:25:17 crc kubenswrapper[4966]: I1217 09:25:17.149816 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" gracePeriod=600 Dec 17 09:25:17 crc kubenswrapper[4966]: E1217 09:25:17.342320 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:25:18 crc kubenswrapper[4966]: I1217 09:25:18.261618 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" exitCode=0 Dec 17 09:25:18 crc kubenswrapper[4966]: I1217 09:25:18.261853 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e"} Dec 17 09:25:18 crc kubenswrapper[4966]: I1217 09:25:18.262468 4966 scope.go:117] "RemoveContainer" containerID="f593595a842f66f60712426dc4265d87e4da7c16085efe1318cfdaeb1f1a07f6" Dec 17 09:25:18 crc kubenswrapper[4966]: I1217 09:25:18.262783 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:25:18 crc kubenswrapper[4966]: E1217 09:25:18.263127 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:25:30 crc kubenswrapper[4966]: I1217 09:25:30.830970 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:25:30 crc kubenswrapper[4966]: E1217 09:25:30.831839 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:25:43 crc kubenswrapper[4966]: I1217 09:25:43.830960 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:25:43 crc kubenswrapper[4966]: E1217 09:25:43.831768 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:25:56 crc kubenswrapper[4966]: I1217 09:25:56.833582 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:25:56 crc kubenswrapper[4966]: E1217 09:25:56.834279 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:26:07 crc kubenswrapper[4966]: I1217 09:26:07.831270 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:26:07 crc kubenswrapper[4966]: E1217 09:26:07.832004 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:26:22 crc kubenswrapper[4966]: I1217 09:26:22.839042 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:26:22 crc kubenswrapper[4966]: E1217 09:26:22.841673 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:26:34 crc kubenswrapper[4966]: I1217 09:26:34.825899 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8j9s7"] Dec 17 09:26:34 crc kubenswrapper[4966]: E1217 09:26:34.829217 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerName="registry-server" Dec 17 09:26:34 crc kubenswrapper[4966]: I1217 09:26:34.829270 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerName="registry-server" Dec 17 09:26:34 crc kubenswrapper[4966]: E1217 09:26:34.829461 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerName="extract-utilities" Dec 17 09:26:34 crc kubenswrapper[4966]: I1217 09:26:34.829474 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerName="extract-utilities" Dec 17 09:26:34 crc kubenswrapper[4966]: E1217 09:26:34.829523 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerName="extract-content" Dec 17 09:26:34 crc kubenswrapper[4966]: I1217 09:26:34.829531 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerName="extract-content" Dec 17 09:26:34 crc kubenswrapper[4966]: I1217 09:26:34.830165 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="63fee3e3-d38a-445c-83c5-91bde266b4b5" containerName="registry-server" Dec 17 09:26:34 crc kubenswrapper[4966]: I1217 09:26:34.834078 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:34 crc kubenswrapper[4966]: I1217 09:26:34.975077 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4n8g\" (UniqueName: \"kubernetes.io/projected/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-kube-api-access-z4n8g\") pod \"community-operators-8j9s7\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:34 crc kubenswrapper[4966]: I1217 09:26:34.975259 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-utilities\") pod \"community-operators-8j9s7\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:34 crc kubenswrapper[4966]: I1217 09:26:34.975289 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-catalog-content\") pod \"community-operators-8j9s7\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:35 crc kubenswrapper[4966]: I1217 09:26:35.077104 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-utilities\") pod \"community-operators-8j9s7\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:35 crc kubenswrapper[4966]: I1217 09:26:35.077733 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-catalog-content\") pod \"community-operators-8j9s7\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:35 crc kubenswrapper[4966]: I1217 09:26:35.077982 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-utilities\") pod \"community-operators-8j9s7\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:35 crc kubenswrapper[4966]: I1217 09:26:35.078051 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-catalog-content\") pod \"community-operators-8j9s7\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:35 crc kubenswrapper[4966]: I1217 09:26:35.078309 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4n8g\" (UniqueName: \"kubernetes.io/projected/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-kube-api-access-z4n8g\") pod \"community-operators-8j9s7\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:35 crc kubenswrapper[4966]: I1217 09:26:35.123735 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4n8g\" (UniqueName: \"kubernetes.io/projected/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-kube-api-access-z4n8g\") pod \"community-operators-8j9s7\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:35 crc kubenswrapper[4966]: I1217 09:26:35.171780 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:35 crc kubenswrapper[4966]: I1217 09:26:35.223576 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8j9s7"] Dec 17 09:26:36 crc kubenswrapper[4966]: I1217 09:26:36.523528 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8j9s7"] Dec 17 09:26:37 crc kubenswrapper[4966]: I1217 09:26:37.400450 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-gvbzk" podUID="fcede706-7499-4c33-b80e-e1a13b7474dd" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 09:26:37 crc kubenswrapper[4966]: I1217 09:26:37.450870 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j9s7" event={"ID":"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe","Type":"ContainerDied","Data":"6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797"} Dec 17 09:26:37 crc kubenswrapper[4966]: I1217 09:26:37.451201 4966 generic.go:334] "Generic (PLEG): container finished" podID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerID="6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797" exitCode=0 Dec 17 09:26:37 crc kubenswrapper[4966]: I1217 09:26:37.451346 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j9s7" event={"ID":"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe","Type":"ContainerStarted","Data":"6fbe3eb201089fd8a2d5334f204cb3a04aca7485ef993ff0e71b713804f98af7"} Dec 17 09:26:37 crc kubenswrapper[4966]: I1217 09:26:37.455057 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 09:26:37 crc kubenswrapper[4966]: I1217 09:26:37.831352 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:26:37 crc kubenswrapper[4966]: E1217 09:26:37.833349 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:26:40 crc kubenswrapper[4966]: I1217 09:26:40.485185 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j9s7" event={"ID":"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe","Type":"ContainerStarted","Data":"a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1"} Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.049917 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" podUID="0be98506-68e8-438f-a46b-2225dd61b871" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.79:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.486444 4966 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-ljxtf container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.487373 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-ljxtf" podUID="50255104-8dc1-485a-b18b-b2f86d9150b4" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.537122 4966 patch_prober.go:28] interesting pod/console-operator-58897d9998-f4bsv container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.537410 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" podUID="b053b6a3-d480-4dee-9c7f-146c7d084e06" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.578069 4966 patch_prober.go:28] interesting pod/console-operator-58897d9998-f4bsv container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.578138 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-f4bsv" podUID="b053b6a3-d480-4dee-9c7f-146c7d084e06" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.803641 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="b562eab8-f0a5-4de7-a251-367555da04bf" containerName="galera" probeResult="failure" output="command timed out" Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.804216 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b562eab8-f0a5-4de7-a251-367555da04bf" containerName="galera" probeResult="failure" output="command timed out" Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.806350 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="ece61b7a-78e0-4e28-9f6d-62c89db6c01d" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.808981 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" podUID="46dfa510-96af-4d45-9be7-5bd9a8588a61" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 09:26:42 crc kubenswrapper[4966]: I1217 09:26:42.812961 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="ece61b7a-78e0-4e28-9f6d-62c89db6c01d" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 17 09:26:43 crc kubenswrapper[4966]: I1217 09:26:43.828654 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="83bca014-fd3a-430d-b532-77c1429a3728" containerName="galera" probeResult="failure" output="command timed out" Dec 17 09:26:43 crc kubenswrapper[4966]: I1217 09:26:43.829331 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="83bca014-fd3a-430d-b532-77c1429a3728" containerName="galera" probeResult="failure" output="command timed out" Dec 17 09:26:46 crc kubenswrapper[4966]: I1217 09:26:46.569923 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j9s7" event={"ID":"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe","Type":"ContainerDied","Data":"a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1"} Dec 17 09:26:46 crc kubenswrapper[4966]: I1217 09:26:46.569909 4966 generic.go:334] "Generic (PLEG): container finished" podID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerID="a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1" exitCode=0 Dec 17 09:26:47 crc kubenswrapper[4966]: I1217 09:26:47.591918 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j9s7" event={"ID":"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe","Type":"ContainerStarted","Data":"5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1"} Dec 17 09:26:47 crc kubenswrapper[4966]: I1217 09:26:47.715097 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8j9s7" podStartSLOduration=3.986443208 podStartE2EDuration="13.71159659s" podCreationTimestamp="2025-12-17 09:26:34 +0000 UTC" firstStartedPulling="2025-12-17 09:26:37.454080517 +0000 UTC m=+3932.999150459" lastFinishedPulling="2025-12-17 09:26:47.179233899 +0000 UTC m=+3942.724303841" observedRunningTime="2025-12-17 09:26:47.704381934 +0000 UTC m=+3943.249451876" watchObservedRunningTime="2025-12-17 09:26:47.71159659 +0000 UTC m=+3943.256666532" Dec 17 09:26:50 crc kubenswrapper[4966]: I1217 09:26:50.836547 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:26:50 crc kubenswrapper[4966]: E1217 09:26:50.837477 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:26:55 crc kubenswrapper[4966]: I1217 09:26:55.173244 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:55 crc kubenswrapper[4966]: I1217 09:26:55.174683 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:26:56 crc kubenswrapper[4966]: I1217 09:26:56.219581 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-8j9s7" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerName="registry-server" probeResult="failure" output=< Dec 17 09:26:56 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:26:56 crc kubenswrapper[4966]: > Dec 17 09:27:03 crc kubenswrapper[4966]: I1217 09:27:03.832721 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:27:03 crc kubenswrapper[4966]: E1217 09:27:03.834405 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:27:05 crc kubenswrapper[4966]: I1217 09:27:05.240927 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:27:05 crc kubenswrapper[4966]: I1217 09:27:05.301023 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:27:05 crc kubenswrapper[4966]: I1217 09:27:05.807101 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8j9s7"] Dec 17 09:27:06 crc kubenswrapper[4966]: I1217 09:27:06.788600 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8j9s7" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerName="registry-server" containerID="cri-o://5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1" gracePeriod=2 Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.772116 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.805897 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j9s7" event={"ID":"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe","Type":"ContainerDied","Data":"5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1"} Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.805920 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8j9s7" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.806376 4966 scope.go:117] "RemoveContainer" containerID="5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.805802 4966 generic.go:334] "Generic (PLEG): container finished" podID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerID="5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1" exitCode=0 Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.806624 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8j9s7" event={"ID":"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe","Type":"ContainerDied","Data":"6fbe3eb201089fd8a2d5334f204cb3a04aca7485ef993ff0e71b713804f98af7"} Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.844105 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4n8g\" (UniqueName: \"kubernetes.io/projected/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-kube-api-access-z4n8g\") pod \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.844229 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-utilities\") pod \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.844277 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-catalog-content\") pod \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\" (UID: \"929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe\") " Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.846366 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-utilities" (OuterVolumeSpecName: "utilities") pod "929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" (UID: "929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.856429 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-kube-api-access-z4n8g" (OuterVolumeSpecName: "kube-api-access-z4n8g") pod "929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" (UID: "929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe"). InnerVolumeSpecName "kube-api-access-z4n8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.869364 4966 scope.go:117] "RemoveContainer" containerID="a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.906350 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" (UID: "929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.912388 4966 scope.go:117] "RemoveContainer" containerID="6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.949609 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4n8g\" (UniqueName: \"kubernetes.io/projected/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-kube-api-access-z4n8g\") on node \"crc\" DevicePath \"\"" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.949822 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.949853 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.963202 4966 scope.go:117] "RemoveContainer" containerID="5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1" Dec 17 09:27:07 crc kubenswrapper[4966]: E1217 09:27:07.964284 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1\": container with ID starting with 5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1 not found: ID does not exist" containerID="5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.964399 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1"} err="failed to get container status \"5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1\": rpc error: code = NotFound desc = could not find container \"5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1\": container with ID starting with 5b5f9e1a17cf1cdc67266cc9f6375b625fbf37a6d87644ba6e0fca0623bdeab1 not found: ID does not exist" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.964497 4966 scope.go:117] "RemoveContainer" containerID="a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1" Dec 17 09:27:07 crc kubenswrapper[4966]: E1217 09:27:07.964854 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1\": container with ID starting with a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1 not found: ID does not exist" containerID="a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.965001 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1"} err="failed to get container status \"a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1\": rpc error: code = NotFound desc = could not find container \"a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1\": container with ID starting with a49f0dec9dd558a5d787086039086d6d39c8900f2ea36a7c27c8ead37bf087c1 not found: ID does not exist" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.965030 4966 scope.go:117] "RemoveContainer" containerID="6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797" Dec 17 09:27:07 crc kubenswrapper[4966]: E1217 09:27:07.965843 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797\": container with ID starting with 6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797 not found: ID does not exist" containerID="6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797" Dec 17 09:27:07 crc kubenswrapper[4966]: I1217 09:27:07.965970 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797"} err="failed to get container status \"6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797\": rpc error: code = NotFound desc = could not find container \"6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797\": container with ID starting with 6746e2db0e5936a7137839cf96f8a41b390660fd0a7f5e0e5aea7d3ac77a7797 not found: ID does not exist" Dec 17 09:27:08 crc kubenswrapper[4966]: I1217 09:27:08.152159 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8j9s7"] Dec 17 09:27:08 crc kubenswrapper[4966]: I1217 09:27:08.161157 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8j9s7"] Dec 17 09:27:08 crc kubenswrapper[4966]: I1217 09:27:08.842380 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" path="/var/lib/kubelet/pods/929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe/volumes" Dec 17 09:27:13 crc kubenswrapper[4966]: E1217 09:27:13.838411 4966 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.008s" Dec 17 09:27:13 crc kubenswrapper[4966]: I1217 09:27:13.849173 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="83bca014-fd3a-430d-b532-77c1429a3728" containerName="galera" probeResult="failure" output="command timed out" Dec 17 09:27:13 crc kubenswrapper[4966]: I1217 09:27:13.857994 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="83bca014-fd3a-430d-b532-77c1429a3728" containerName="galera" probeResult="failure" output="command timed out" Dec 17 09:27:14 crc kubenswrapper[4966]: I1217 09:27:14.832503 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:27:14 crc kubenswrapper[4966]: E1217 09:27:14.834158 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:27:28 crc kubenswrapper[4966]: I1217 09:27:28.833256 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:27:28 crc kubenswrapper[4966]: E1217 09:27:28.834849 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:27:40 crc kubenswrapper[4966]: I1217 09:27:40.831339 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:27:40 crc kubenswrapper[4966]: E1217 09:27:40.832383 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:27:52 crc kubenswrapper[4966]: I1217 09:27:52.832581 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:27:52 crc kubenswrapper[4966]: E1217 09:27:52.833720 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:28:03 crc kubenswrapper[4966]: I1217 09:28:03.831211 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:28:03 crc kubenswrapper[4966]: E1217 09:28:03.832149 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:28:16 crc kubenswrapper[4966]: I1217 09:28:16.830987 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:28:16 crc kubenswrapper[4966]: E1217 09:28:16.833487 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:28:31 crc kubenswrapper[4966]: I1217 09:28:31.830357 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:28:31 crc kubenswrapper[4966]: E1217 09:28:31.831133 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.130070 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qp4nv"] Dec 17 09:28:33 crc kubenswrapper[4966]: E1217 09:28:33.132148 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerName="registry-server" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.132174 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerName="registry-server" Dec 17 09:28:33 crc kubenswrapper[4966]: E1217 09:28:33.132191 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerName="extract-utilities" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.132197 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerName="extract-utilities" Dec 17 09:28:33 crc kubenswrapper[4966]: E1217 09:28:33.132224 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerName="extract-content" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.132231 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerName="extract-content" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.132888 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="929545b0-1d82-4bb5-a5f7-c8fc04bc3ebe" containerName="registry-server" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.136515 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.183638 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzq6h\" (UniqueName: \"kubernetes.io/projected/d388c5f9-d9ef-497b-869e-581e308799e6-kube-api-access-kzq6h\") pod \"redhat-operators-qp4nv\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.183727 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-utilities\") pod \"redhat-operators-qp4nv\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.184043 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-catalog-content\") pod \"redhat-operators-qp4nv\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.281013 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qp4nv"] Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.285858 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-catalog-content\") pod \"redhat-operators-qp4nv\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.286025 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzq6h\" (UniqueName: \"kubernetes.io/projected/d388c5f9-d9ef-497b-869e-581e308799e6-kube-api-access-kzq6h\") pod \"redhat-operators-qp4nv\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.286059 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-utilities\") pod \"redhat-operators-qp4nv\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.289105 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-utilities\") pod \"redhat-operators-qp4nv\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.289760 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-catalog-content\") pod \"redhat-operators-qp4nv\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.398371 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzq6h\" (UniqueName: \"kubernetes.io/projected/d388c5f9-d9ef-497b-869e-581e308799e6-kube-api-access-kzq6h\") pod \"redhat-operators-qp4nv\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:33 crc kubenswrapper[4966]: I1217 09:28:33.460640 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:34 crc kubenswrapper[4966]: I1217 09:28:34.429016 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qp4nv"] Dec 17 09:28:34 crc kubenswrapper[4966]: W1217 09:28:34.441304 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd388c5f9_d9ef_497b_869e_581e308799e6.slice/crio-f53118440814d3aae8e815a6ed407033d87383fc3f5d84518a1204be58b12f4d WatchSource:0}: Error finding container f53118440814d3aae8e815a6ed407033d87383fc3f5d84518a1204be58b12f4d: Status 404 returned error can't find the container with id f53118440814d3aae8e815a6ed407033d87383fc3f5d84518a1204be58b12f4d Dec 17 09:28:34 crc kubenswrapper[4966]: I1217 09:28:34.697802 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qp4nv" event={"ID":"d388c5f9-d9ef-497b-869e-581e308799e6","Type":"ContainerStarted","Data":"eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786"} Dec 17 09:28:34 crc kubenswrapper[4966]: I1217 09:28:34.698134 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qp4nv" event={"ID":"d388c5f9-d9ef-497b-869e-581e308799e6","Type":"ContainerStarted","Data":"f53118440814d3aae8e815a6ed407033d87383fc3f5d84518a1204be58b12f4d"} Dec 17 09:28:35 crc kubenswrapper[4966]: I1217 09:28:35.707146 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qp4nv" event={"ID":"d388c5f9-d9ef-497b-869e-581e308799e6","Type":"ContainerDied","Data":"eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786"} Dec 17 09:28:35 crc kubenswrapper[4966]: I1217 09:28:35.707232 4966 generic.go:334] "Generic (PLEG): container finished" podID="d388c5f9-d9ef-497b-869e-581e308799e6" containerID="eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786" exitCode=0 Dec 17 09:28:38 crc kubenswrapper[4966]: I1217 09:28:38.734420 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qp4nv" event={"ID":"d388c5f9-d9ef-497b-869e-581e308799e6","Type":"ContainerStarted","Data":"1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a"} Dec 17 09:28:42 crc kubenswrapper[4966]: I1217 09:28:42.772052 4966 generic.go:334] "Generic (PLEG): container finished" podID="d388c5f9-d9ef-497b-869e-581e308799e6" containerID="1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a" exitCode=0 Dec 17 09:28:42 crc kubenswrapper[4966]: I1217 09:28:42.772105 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qp4nv" event={"ID":"d388c5f9-d9ef-497b-869e-581e308799e6","Type":"ContainerDied","Data":"1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a"} Dec 17 09:28:43 crc kubenswrapper[4966]: I1217 09:28:43.836032 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:28:43 crc kubenswrapper[4966]: E1217 09:28:43.844049 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:28:44 crc kubenswrapper[4966]: I1217 09:28:44.793630 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qp4nv" event={"ID":"d388c5f9-d9ef-497b-869e-581e308799e6","Type":"ContainerStarted","Data":"21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c"} Dec 17 09:28:44 crc kubenswrapper[4966]: I1217 09:28:44.815114 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qp4nv" podStartSLOduration=3.953805187 podStartE2EDuration="11.814807279s" podCreationTimestamp="2025-12-17 09:28:33 +0000 UTC" firstStartedPulling="2025-12-17 09:28:35.709757503 +0000 UTC m=+4051.254827445" lastFinishedPulling="2025-12-17 09:28:43.570759595 +0000 UTC m=+4059.115829537" observedRunningTime="2025-12-17 09:28:44.810659156 +0000 UTC m=+4060.355729098" watchObservedRunningTime="2025-12-17 09:28:44.814807279 +0000 UTC m=+4060.359877211" Dec 17 09:28:53 crc kubenswrapper[4966]: I1217 09:28:53.461561 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:53 crc kubenswrapper[4966]: I1217 09:28:53.462342 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:28:54 crc kubenswrapper[4966]: I1217 09:28:54.509388 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qp4nv" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" containerName="registry-server" probeResult="failure" output=< Dec 17 09:28:54 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:28:54 crc kubenswrapper[4966]: > Dec 17 09:28:55 crc kubenswrapper[4966]: I1217 09:28:55.830689 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:28:55 crc kubenswrapper[4966]: E1217 09:28:55.831894 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:29:03 crc kubenswrapper[4966]: I1217 09:29:03.516012 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:29:03 crc kubenswrapper[4966]: I1217 09:29:03.576602 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:29:04 crc kubenswrapper[4966]: I1217 09:29:04.344289 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qp4nv"] Dec 17 09:29:05 crc kubenswrapper[4966]: I1217 09:29:05.021923 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qp4nv" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" containerName="registry-server" containerID="cri-o://21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c" gracePeriod=2 Dec 17 09:29:05 crc kubenswrapper[4966]: I1217 09:29:05.913470 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:29:05 crc kubenswrapper[4966]: I1217 09:29:05.949787 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzq6h\" (UniqueName: \"kubernetes.io/projected/d388c5f9-d9ef-497b-869e-581e308799e6-kube-api-access-kzq6h\") pod \"d388c5f9-d9ef-497b-869e-581e308799e6\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " Dec 17 09:29:05 crc kubenswrapper[4966]: I1217 09:29:05.950079 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-catalog-content\") pod \"d388c5f9-d9ef-497b-869e-581e308799e6\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " Dec 17 09:29:05 crc kubenswrapper[4966]: I1217 09:29:05.950156 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-utilities\") pod \"d388c5f9-d9ef-497b-869e-581e308799e6\" (UID: \"d388c5f9-d9ef-497b-869e-581e308799e6\") " Dec 17 09:29:05 crc kubenswrapper[4966]: I1217 09:29:05.952057 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-utilities" (OuterVolumeSpecName: "utilities") pod "d388c5f9-d9ef-497b-869e-581e308799e6" (UID: "d388c5f9-d9ef-497b-869e-581e308799e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:29:05 crc kubenswrapper[4966]: I1217 09:29:05.970286 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d388c5f9-d9ef-497b-869e-581e308799e6-kube-api-access-kzq6h" (OuterVolumeSpecName: "kube-api-access-kzq6h") pod "d388c5f9-d9ef-497b-869e-581e308799e6" (UID: "d388c5f9-d9ef-497b-869e-581e308799e6"). InnerVolumeSpecName "kube-api-access-kzq6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.031212 4966 generic.go:334] "Generic (PLEG): container finished" podID="d388c5f9-d9ef-497b-869e-581e308799e6" containerID="21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c" exitCode=0 Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.031293 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qp4nv" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.031319 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qp4nv" event={"ID":"d388c5f9-d9ef-497b-869e-581e308799e6","Type":"ContainerDied","Data":"21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c"} Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.033392 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qp4nv" event={"ID":"d388c5f9-d9ef-497b-869e-581e308799e6","Type":"ContainerDied","Data":"f53118440814d3aae8e815a6ed407033d87383fc3f5d84518a1204be58b12f4d"} Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.033504 4966 scope.go:117] "RemoveContainer" containerID="21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.052623 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.052659 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzq6h\" (UniqueName: \"kubernetes.io/projected/d388c5f9-d9ef-497b-869e-581e308799e6-kube-api-access-kzq6h\") on node \"crc\" DevicePath \"\"" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.069526 4966 scope.go:117] "RemoveContainer" containerID="1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.085519 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d388c5f9-d9ef-497b-869e-581e308799e6" (UID: "d388c5f9-d9ef-497b-869e-581e308799e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.096516 4966 scope.go:117] "RemoveContainer" containerID="eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.139789 4966 scope.go:117] "RemoveContainer" containerID="21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c" Dec 17 09:29:06 crc kubenswrapper[4966]: E1217 09:29:06.140751 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c\": container with ID starting with 21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c not found: ID does not exist" containerID="21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.140933 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c"} err="failed to get container status \"21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c\": rpc error: code = NotFound desc = could not find container \"21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c\": container with ID starting with 21c4047ce737be7845c471ebb18b793406dda1d64519a879afa06abbc76e905c not found: ID does not exist" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.140968 4966 scope.go:117] "RemoveContainer" containerID="1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a" Dec 17 09:29:06 crc kubenswrapper[4966]: E1217 09:29:06.141400 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a\": container with ID starting with 1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a not found: ID does not exist" containerID="1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.141481 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a"} err="failed to get container status \"1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a\": rpc error: code = NotFound desc = could not find container \"1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a\": container with ID starting with 1e549990043f147425a66eb948498d83ae5762318975652f4305bc4a945ba95a not found: ID does not exist" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.141533 4966 scope.go:117] "RemoveContainer" containerID="eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786" Dec 17 09:29:06 crc kubenswrapper[4966]: E1217 09:29:06.141899 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786\": container with ID starting with eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786 not found: ID does not exist" containerID="eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.141923 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786"} err="failed to get container status \"eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786\": rpc error: code = NotFound desc = could not find container \"eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786\": container with ID starting with eda1940b99545bcaba3f6ca7b5e0726cb762da306f9cb7627d062deda97e1786 not found: ID does not exist" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.154641 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d388c5f9-d9ef-497b-869e-581e308799e6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.383156 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qp4nv"] Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.396176 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qp4nv"] Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.832821 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:29:06 crc kubenswrapper[4966]: E1217 09:29:06.833713 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:29:06 crc kubenswrapper[4966]: I1217 09:29:06.846838 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" path="/var/lib/kubelet/pods/d388c5f9-d9ef-497b-869e-581e308799e6/volumes" Dec 17 09:29:17 crc kubenswrapper[4966]: I1217 09:29:17.831041 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:29:17 crc kubenswrapper[4966]: E1217 09:29:17.834749 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:29:30 crc kubenswrapper[4966]: I1217 09:29:30.830783 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:29:30 crc kubenswrapper[4966]: E1217 09:29:30.831829 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:29:41 crc kubenswrapper[4966]: I1217 09:29:41.831254 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:29:41 crc kubenswrapper[4966]: E1217 09:29:41.832053 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:29:56 crc kubenswrapper[4966]: I1217 09:29:56.831072 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:29:56 crc kubenswrapper[4966]: E1217 09:29:56.831768 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.401701 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj"] Dec 17 09:30:00 crc kubenswrapper[4966]: E1217 09:30:00.406344 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" containerName="extract-utilities" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.406390 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" containerName="extract-utilities" Dec 17 09:30:00 crc kubenswrapper[4966]: E1217 09:30:00.406411 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" containerName="registry-server" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.406421 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" containerName="registry-server" Dec 17 09:30:00 crc kubenswrapper[4966]: E1217 09:30:00.406441 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" containerName="extract-content" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.406447 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" containerName="extract-content" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.406934 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="d388c5f9-d9ef-497b-869e-581e308799e6" containerName="registry-server" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.409039 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.414292 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.423306 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.425276 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj"] Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.578262 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsllr\" (UniqueName: \"kubernetes.io/projected/09380640-3b0e-4f60-bfa0-692b6d785749-kube-api-access-dsllr\") pod \"collect-profiles-29432730-8jdsj\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.578622 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09380640-3b0e-4f60-bfa0-692b6d785749-secret-volume\") pod \"collect-profiles-29432730-8jdsj\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.578748 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09380640-3b0e-4f60-bfa0-692b6d785749-config-volume\") pod \"collect-profiles-29432730-8jdsj\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.681031 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09380640-3b0e-4f60-bfa0-692b6d785749-secret-volume\") pod \"collect-profiles-29432730-8jdsj\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.681177 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09380640-3b0e-4f60-bfa0-692b6d785749-config-volume\") pod \"collect-profiles-29432730-8jdsj\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.681289 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsllr\" (UniqueName: \"kubernetes.io/projected/09380640-3b0e-4f60-bfa0-692b6d785749-kube-api-access-dsllr\") pod \"collect-profiles-29432730-8jdsj\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.685574 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09380640-3b0e-4f60-bfa0-692b6d785749-config-volume\") pod \"collect-profiles-29432730-8jdsj\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.700140 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09380640-3b0e-4f60-bfa0-692b6d785749-secret-volume\") pod \"collect-profiles-29432730-8jdsj\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.706624 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsllr\" (UniqueName: \"kubernetes.io/projected/09380640-3b0e-4f60-bfa0-692b6d785749-kube-api-access-dsllr\") pod \"collect-profiles-29432730-8jdsj\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:00 crc kubenswrapper[4966]: I1217 09:30:00.741192 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:01 crc kubenswrapper[4966]: I1217 09:30:01.520571 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj"] Dec 17 09:30:01 crc kubenswrapper[4966]: I1217 09:30:01.887473 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" event={"ID":"09380640-3b0e-4f60-bfa0-692b6d785749","Type":"ContainerStarted","Data":"7da1763574690e772b658e0c7b0f97af8b6b89ea00632b82a77139e4aca6934a"} Dec 17 09:30:02 crc kubenswrapper[4966]: I1217 09:30:02.898332 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" event={"ID":"09380640-3b0e-4f60-bfa0-692b6d785749","Type":"ContainerDied","Data":"fbb5dbbd73d6d9a85f7e16532d430b24206f1a9b18bde959083b686eeb0d2b22"} Dec 17 09:30:02 crc kubenswrapper[4966]: I1217 09:30:02.898207 4966 generic.go:334] "Generic (PLEG): container finished" podID="09380640-3b0e-4f60-bfa0-692b6d785749" containerID="fbb5dbbd73d6d9a85f7e16532d430b24206f1a9b18bde959083b686eeb0d2b22" exitCode=0 Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.320490 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.460911 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsllr\" (UniqueName: \"kubernetes.io/projected/09380640-3b0e-4f60-bfa0-692b6d785749-kube-api-access-dsllr\") pod \"09380640-3b0e-4f60-bfa0-692b6d785749\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.461291 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09380640-3b0e-4f60-bfa0-692b6d785749-secret-volume\") pod \"09380640-3b0e-4f60-bfa0-692b6d785749\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.461389 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09380640-3b0e-4f60-bfa0-692b6d785749-config-volume\") pod \"09380640-3b0e-4f60-bfa0-692b6d785749\" (UID: \"09380640-3b0e-4f60-bfa0-692b6d785749\") " Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.463201 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09380640-3b0e-4f60-bfa0-692b6d785749-config-volume" (OuterVolumeSpecName: "config-volume") pod "09380640-3b0e-4f60-bfa0-692b6d785749" (UID: "09380640-3b0e-4f60-bfa0-692b6d785749"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.477654 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09380640-3b0e-4f60-bfa0-692b6d785749-kube-api-access-dsllr" (OuterVolumeSpecName: "kube-api-access-dsllr") pod "09380640-3b0e-4f60-bfa0-692b6d785749" (UID: "09380640-3b0e-4f60-bfa0-692b6d785749"). InnerVolumeSpecName "kube-api-access-dsllr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.489521 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09380640-3b0e-4f60-bfa0-692b6d785749-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "09380640-3b0e-4f60-bfa0-692b6d785749" (UID: "09380640-3b0e-4f60-bfa0-692b6d785749"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.564589 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsllr\" (UniqueName: \"kubernetes.io/projected/09380640-3b0e-4f60-bfa0-692b6d785749-kube-api-access-dsllr\") on node \"crc\" DevicePath \"\"" Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.564633 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09380640-3b0e-4f60-bfa0-692b6d785749-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.564645 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09380640-3b0e-4f60-bfa0-692b6d785749-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.918432 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" event={"ID":"09380640-3b0e-4f60-bfa0-692b6d785749","Type":"ContainerDied","Data":"7da1763574690e772b658e0c7b0f97af8b6b89ea00632b82a77139e4aca6934a"} Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.918483 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj" Dec 17 09:30:04 crc kubenswrapper[4966]: I1217 09:30:04.918501 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7da1763574690e772b658e0c7b0f97af8b6b89ea00632b82a77139e4aca6934a" Dec 17 09:30:05 crc kubenswrapper[4966]: E1217 09:30:05.059361 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09380640_3b0e_4f60_bfa0_692b6d785749.slice/crio-7da1763574690e772b658e0c7b0f97af8b6b89ea00632b82a77139e4aca6934a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09380640_3b0e_4f60_bfa0_692b6d785749.slice\": RecentStats: unable to find data in memory cache]" Dec 17 09:30:05 crc kubenswrapper[4966]: I1217 09:30:05.412594 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8"] Dec 17 09:30:05 crc kubenswrapper[4966]: I1217 09:30:05.422000 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432685-7kjd8"] Dec 17 09:30:06 crc kubenswrapper[4966]: I1217 09:30:06.855066 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1" path="/var/lib/kubelet/pods/4b1a4d2c-b220-4bdb-8d2d-239b4e9f28f1/volumes" Dec 17 09:30:11 crc kubenswrapper[4966]: I1217 09:30:11.831071 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:30:11 crc kubenswrapper[4966]: E1217 09:30:11.831639 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:30:26 crc kubenswrapper[4966]: I1217 09:30:26.831503 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:30:27 crc kubenswrapper[4966]: I1217 09:30:27.119283 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"4e5578a74a8e5e7a291125ba7509b12de429ea30e5bd8aebceb023dc7cc3a7f9"} Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.696777 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr8z"] Dec 17 09:30:44 crc kubenswrapper[4966]: E1217 09:30:44.701624 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09380640-3b0e-4f60-bfa0-692b6d785749" containerName="collect-profiles" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.701669 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="09380640-3b0e-4f60-bfa0-692b6d785749" containerName="collect-profiles" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.703170 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="09380640-3b0e-4f60-bfa0-692b6d785749" containerName="collect-profiles" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.707344 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.799410 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw9rt\" (UniqueName: \"kubernetes.io/projected/091e4ab4-1625-45c3-ace0-b9b15d56ad32-kube-api-access-pw9rt\") pod \"redhat-marketplace-bwr8z\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.799495 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-catalog-content\") pod \"redhat-marketplace-bwr8z\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.799569 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-utilities\") pod \"redhat-marketplace-bwr8z\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.901692 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-catalog-content\") pod \"redhat-marketplace-bwr8z\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.901812 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-utilities\") pod \"redhat-marketplace-bwr8z\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.901982 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw9rt\" (UniqueName: \"kubernetes.io/projected/091e4ab4-1625-45c3-ace0-b9b15d56ad32-kube-api-access-pw9rt\") pod \"redhat-marketplace-bwr8z\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.904287 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-utilities\") pod \"redhat-marketplace-bwr8z\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.904526 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-catalog-content\") pod \"redhat-marketplace-bwr8z\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:44 crc kubenswrapper[4966]: I1217 09:30:44.932627 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw9rt\" (UniqueName: \"kubernetes.io/projected/091e4ab4-1625-45c3-ace0-b9b15d56ad32-kube-api-access-pw9rt\") pod \"redhat-marketplace-bwr8z\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:45 crc kubenswrapper[4966]: I1217 09:30:45.003359 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr8z"] Dec 17 09:30:45 crc kubenswrapper[4966]: I1217 09:30:45.035662 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:46 crc kubenswrapper[4966]: I1217 09:30:46.385826 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr8z"] Dec 17 09:30:47 crc kubenswrapper[4966]: I1217 09:30:47.290453 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr8z" event={"ID":"091e4ab4-1625-45c3-ace0-b9b15d56ad32","Type":"ContainerDied","Data":"98ecc1a4bc815722870da259cfc8722cec65bb3e742617f9b837385bbe75d923"} Dec 17 09:30:47 crc kubenswrapper[4966]: I1217 09:30:47.290524 4966 generic.go:334] "Generic (PLEG): container finished" podID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerID="98ecc1a4bc815722870da259cfc8722cec65bb3e742617f9b837385bbe75d923" exitCode=0 Dec 17 09:30:47 crc kubenswrapper[4966]: I1217 09:30:47.290989 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr8z" event={"ID":"091e4ab4-1625-45c3-ace0-b9b15d56ad32","Type":"ContainerStarted","Data":"6dc8b39d316df6d881f1be6e27944dce3a9e0408af2d809aaf90b8974e397d87"} Dec 17 09:30:49 crc kubenswrapper[4966]: I1217 09:30:49.330431 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr8z" event={"ID":"091e4ab4-1625-45c3-ace0-b9b15d56ad32","Type":"ContainerStarted","Data":"2525a481c129703fef607e61bc651138affd2ce27b75c3dd3fb597f3eb26f63f"} Dec 17 09:30:50 crc kubenswrapper[4966]: I1217 09:30:50.347793 4966 generic.go:334] "Generic (PLEG): container finished" podID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerID="2525a481c129703fef607e61bc651138affd2ce27b75c3dd3fb597f3eb26f63f" exitCode=0 Dec 17 09:30:50 crc kubenswrapper[4966]: I1217 09:30:50.348045 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr8z" event={"ID":"091e4ab4-1625-45c3-ace0-b9b15d56ad32","Type":"ContainerDied","Data":"2525a481c129703fef607e61bc651138affd2ce27b75c3dd3fb597f3eb26f63f"} Dec 17 09:30:51 crc kubenswrapper[4966]: I1217 09:30:51.363841 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr8z" event={"ID":"091e4ab4-1625-45c3-ace0-b9b15d56ad32","Type":"ContainerStarted","Data":"0623f7f2cfc4a91694eba5c071b5b435081a0706026b8d52a8084b41e7b1ed03"} Dec 17 09:30:51 crc kubenswrapper[4966]: I1217 09:30:51.388680 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bwr8z" podStartSLOduration=3.637617263 podStartE2EDuration="7.38774266s" podCreationTimestamp="2025-12-17 09:30:44 +0000 UTC" firstStartedPulling="2025-12-17 09:30:47.292175711 +0000 UTC m=+4182.837245653" lastFinishedPulling="2025-12-17 09:30:51.042301108 +0000 UTC m=+4186.587371050" observedRunningTime="2025-12-17 09:30:51.38372638 +0000 UTC m=+4186.928796322" watchObservedRunningTime="2025-12-17 09:30:51.38774266 +0000 UTC m=+4186.932812602" Dec 17 09:30:55 crc kubenswrapper[4966]: I1217 09:30:55.036794 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:55 crc kubenswrapper[4966]: I1217 09:30:55.038372 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:55 crc kubenswrapper[4966]: I1217 09:30:55.091079 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:56 crc kubenswrapper[4966]: I1217 09:30:56.490981 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:30:59 crc kubenswrapper[4966]: I1217 09:30:59.044966 4966 scope.go:117] "RemoveContainer" containerID="45421059253cac7584638191c6488de7cfdc0673f06d2337bddd34b4dd676cfd" Dec 17 09:30:59 crc kubenswrapper[4966]: I1217 09:30:59.736652 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr8z"] Dec 17 09:30:59 crc kubenswrapper[4966]: I1217 09:30:59.737937 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bwr8z" podUID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerName="registry-server" containerID="cri-o://0623f7f2cfc4a91694eba5c071b5b435081a0706026b8d52a8084b41e7b1ed03" gracePeriod=2 Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.485216 4966 generic.go:334] "Generic (PLEG): container finished" podID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerID="0623f7f2cfc4a91694eba5c071b5b435081a0706026b8d52a8084b41e7b1ed03" exitCode=0 Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.485292 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr8z" event={"ID":"091e4ab4-1625-45c3-ace0-b9b15d56ad32","Type":"ContainerDied","Data":"0623f7f2cfc4a91694eba5c071b5b435081a0706026b8d52a8084b41e7b1ed03"} Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.693227 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.778234 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-catalog-content\") pod \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.778362 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw9rt\" (UniqueName: \"kubernetes.io/projected/091e4ab4-1625-45c3-ace0-b9b15d56ad32-kube-api-access-pw9rt\") pod \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.788786 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091e4ab4-1625-45c3-ace0-b9b15d56ad32-kube-api-access-pw9rt" (OuterVolumeSpecName: "kube-api-access-pw9rt") pod "091e4ab4-1625-45c3-ace0-b9b15d56ad32" (UID: "091e4ab4-1625-45c3-ace0-b9b15d56ad32"). InnerVolumeSpecName "kube-api-access-pw9rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.807782 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "091e4ab4-1625-45c3-ace0-b9b15d56ad32" (UID: "091e4ab4-1625-45c3-ace0-b9b15d56ad32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.879751 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-utilities\") pod \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\" (UID: \"091e4ab4-1625-45c3-ace0-b9b15d56ad32\") " Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.880831 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-utilities" (OuterVolumeSpecName: "utilities") pod "091e4ab4-1625-45c3-ace0-b9b15d56ad32" (UID: "091e4ab4-1625-45c3-ace0-b9b15d56ad32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.881818 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.881847 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw9rt\" (UniqueName: \"kubernetes.io/projected/091e4ab4-1625-45c3-ace0-b9b15d56ad32-kube-api-access-pw9rt\") on node \"crc\" DevicePath \"\"" Dec 17 09:31:00 crc kubenswrapper[4966]: I1217 09:31:00.881858 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/091e4ab4-1625-45c3-ace0-b9b15d56ad32-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:31:01 crc kubenswrapper[4966]: I1217 09:31:01.496828 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bwr8z" event={"ID":"091e4ab4-1625-45c3-ace0-b9b15d56ad32","Type":"ContainerDied","Data":"6dc8b39d316df6d881f1be6e27944dce3a9e0408af2d809aaf90b8974e397d87"} Dec 17 09:31:01 crc kubenswrapper[4966]: I1217 09:31:01.496903 4966 scope.go:117] "RemoveContainer" containerID="0623f7f2cfc4a91694eba5c071b5b435081a0706026b8d52a8084b41e7b1ed03" Dec 17 09:31:01 crc kubenswrapper[4966]: I1217 09:31:01.496933 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bwr8z" Dec 17 09:31:01 crc kubenswrapper[4966]: I1217 09:31:01.533268 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr8z"] Dec 17 09:31:01 crc kubenswrapper[4966]: I1217 09:31:01.542312 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bwr8z"] Dec 17 09:31:01 crc kubenswrapper[4966]: I1217 09:31:01.551815 4966 scope.go:117] "RemoveContainer" containerID="2525a481c129703fef607e61bc651138affd2ce27b75c3dd3fb597f3eb26f63f" Dec 17 09:31:01 crc kubenswrapper[4966]: I1217 09:31:01.581515 4966 scope.go:117] "RemoveContainer" containerID="98ecc1a4bc815722870da259cfc8722cec65bb3e742617f9b837385bbe75d923" Dec 17 09:31:02 crc kubenswrapper[4966]: I1217 09:31:02.844720 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" path="/var/lib/kubelet/pods/091e4ab4-1625-45c3-ace0-b9b15d56ad32/volumes" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.155596 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-thtlz"] Dec 17 09:31:38 crc kubenswrapper[4966]: E1217 09:31:38.159631 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerName="extract-content" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.159684 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerName="extract-content" Dec 17 09:31:38 crc kubenswrapper[4966]: E1217 09:31:38.159717 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerName="registry-server" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.159726 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerName="registry-server" Dec 17 09:31:38 crc kubenswrapper[4966]: E1217 09:31:38.159748 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerName="extract-utilities" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.159757 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerName="extract-utilities" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.160564 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="091e4ab4-1625-45c3-ace0-b9b15d56ad32" containerName="registry-server" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.165151 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.182382 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-thtlz"] Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.263390 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-catalog-content\") pod \"certified-operators-thtlz\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.263604 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh526\" (UniqueName: \"kubernetes.io/projected/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-kube-api-access-sh526\") pod \"certified-operators-thtlz\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.263700 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-utilities\") pod \"certified-operators-thtlz\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.365013 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-catalog-content\") pod \"certified-operators-thtlz\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.365120 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh526\" (UniqueName: \"kubernetes.io/projected/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-kube-api-access-sh526\") pod \"certified-operators-thtlz\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.365191 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-utilities\") pod \"certified-operators-thtlz\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.366844 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-catalog-content\") pod \"certified-operators-thtlz\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.367263 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-utilities\") pod \"certified-operators-thtlz\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.398282 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh526\" (UniqueName: \"kubernetes.io/projected/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-kube-api-access-sh526\") pod \"certified-operators-thtlz\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:38 crc kubenswrapper[4966]: I1217 09:31:38.491352 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:39 crc kubenswrapper[4966]: W1217 09:31:39.231125 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2ad4f7e_ecb3_4941_aa7b_70568f0fd15a.slice/crio-d3d45b40c630c7bf9524d32f7514f7832d02c031a4c517f9fec9d7b5339323a4 WatchSource:0}: Error finding container d3d45b40c630c7bf9524d32f7514f7832d02c031a4c517f9fec9d7b5339323a4: Status 404 returned error can't find the container with id d3d45b40c630c7bf9524d32f7514f7832d02c031a4c517f9fec9d7b5339323a4 Dec 17 09:31:39 crc kubenswrapper[4966]: I1217 09:31:39.239699 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-thtlz"] Dec 17 09:31:39 crc kubenswrapper[4966]: I1217 09:31:39.867517 4966 generic.go:334] "Generic (PLEG): container finished" podID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerID="23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1" exitCode=0 Dec 17 09:31:39 crc kubenswrapper[4966]: I1217 09:31:39.868005 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-thtlz" event={"ID":"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a","Type":"ContainerDied","Data":"23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1"} Dec 17 09:31:39 crc kubenswrapper[4966]: I1217 09:31:39.868060 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-thtlz" event={"ID":"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a","Type":"ContainerStarted","Data":"d3d45b40c630c7bf9524d32f7514f7832d02c031a4c517f9fec9d7b5339323a4"} Dec 17 09:31:39 crc kubenswrapper[4966]: I1217 09:31:39.871834 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 09:31:41 crc kubenswrapper[4966]: I1217 09:31:41.898409 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-thtlz" event={"ID":"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a","Type":"ContainerStarted","Data":"20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553"} Dec 17 09:31:42 crc kubenswrapper[4966]: I1217 09:31:42.911183 4966 generic.go:334] "Generic (PLEG): container finished" podID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerID="20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553" exitCode=0 Dec 17 09:31:42 crc kubenswrapper[4966]: I1217 09:31:42.911269 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-thtlz" event={"ID":"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a","Type":"ContainerDied","Data":"20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553"} Dec 17 09:31:43 crc kubenswrapper[4966]: I1217 09:31:43.923890 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-thtlz" event={"ID":"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a","Type":"ContainerStarted","Data":"87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e"} Dec 17 09:31:43 crc kubenswrapper[4966]: I1217 09:31:43.947819 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-thtlz" podStartSLOduration=2.494447978 podStartE2EDuration="5.947280244s" podCreationTimestamp="2025-12-17 09:31:38 +0000 UTC" firstStartedPulling="2025-12-17 09:31:39.869938393 +0000 UTC m=+4235.415008345" lastFinishedPulling="2025-12-17 09:31:43.322770669 +0000 UTC m=+4238.867840611" observedRunningTime="2025-12-17 09:31:43.943421569 +0000 UTC m=+4239.488491521" watchObservedRunningTime="2025-12-17 09:31:43.947280244 +0000 UTC m=+4239.492350186" Dec 17 09:31:48 crc kubenswrapper[4966]: I1217 09:31:48.493313 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:48 crc kubenswrapper[4966]: I1217 09:31:48.493845 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:49 crc kubenswrapper[4966]: I1217 09:31:49.541085 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-thtlz" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerName="registry-server" probeResult="failure" output=< Dec 17 09:31:49 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:31:49 crc kubenswrapper[4966]: > Dec 17 09:31:58 crc kubenswrapper[4966]: I1217 09:31:58.561255 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:58 crc kubenswrapper[4966]: I1217 09:31:58.619656 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:31:58 crc kubenswrapper[4966]: I1217 09:31:58.807729 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-thtlz"] Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.071526 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-thtlz" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerName="registry-server" containerID="cri-o://87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e" gracePeriod=2 Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.705991 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.860276 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-utilities\") pod \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.860415 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh526\" (UniqueName: \"kubernetes.io/projected/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-kube-api-access-sh526\") pod \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.860510 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-catalog-content\") pod \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\" (UID: \"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a\") " Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.861460 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-utilities" (OuterVolumeSpecName: "utilities") pod "a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" (UID: "a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.868975 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-kube-api-access-sh526" (OuterVolumeSpecName: "kube-api-access-sh526") pod "a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" (UID: "a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a"). InnerVolumeSpecName "kube-api-access-sh526". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.938057 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" (UID: "a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.962612 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh526\" (UniqueName: \"kubernetes.io/projected/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-kube-api-access-sh526\") on node \"crc\" DevicePath \"\"" Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.962660 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:32:00 crc kubenswrapper[4966]: I1217 09:32:00.962672 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.080165 4966 generic.go:334] "Generic (PLEG): container finished" podID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerID="87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e" exitCode=0 Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.080215 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-thtlz" event={"ID":"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a","Type":"ContainerDied","Data":"87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e"} Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.080245 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-thtlz" event={"ID":"a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a","Type":"ContainerDied","Data":"d3d45b40c630c7bf9524d32f7514f7832d02c031a4c517f9fec9d7b5339323a4"} Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.080297 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-thtlz" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.080821 4966 scope.go:117] "RemoveContainer" containerID="87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.112061 4966 scope.go:117] "RemoveContainer" containerID="20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.124514 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-thtlz"] Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.135543 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-thtlz"] Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.138615 4966 scope.go:117] "RemoveContainer" containerID="23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.176384 4966 scope.go:117] "RemoveContainer" containerID="87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e" Dec 17 09:32:01 crc kubenswrapper[4966]: E1217 09:32:01.177371 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e\": container with ID starting with 87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e not found: ID does not exist" containerID="87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.177541 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e"} err="failed to get container status \"87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e\": rpc error: code = NotFound desc = could not find container \"87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e\": container with ID starting with 87eb2d67f4d4fbf0f8b89c8a7166320eb9bc2d0c210e973ebfde73590d8bf48e not found: ID does not exist" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.177571 4966 scope.go:117] "RemoveContainer" containerID="20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553" Dec 17 09:32:01 crc kubenswrapper[4966]: E1217 09:32:01.177987 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553\": container with ID starting with 20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553 not found: ID does not exist" containerID="20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.178020 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553"} err="failed to get container status \"20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553\": rpc error: code = NotFound desc = could not find container \"20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553\": container with ID starting with 20dcea788791854592a3a524013673f470c88e73d5d101e0ab57706e05627553 not found: ID does not exist" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.178038 4966 scope.go:117] "RemoveContainer" containerID="23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1" Dec 17 09:32:01 crc kubenswrapper[4966]: E1217 09:32:01.178251 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1\": container with ID starting with 23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1 not found: ID does not exist" containerID="23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1" Dec 17 09:32:01 crc kubenswrapper[4966]: I1217 09:32:01.178270 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1"} err="failed to get container status \"23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1\": rpc error: code = NotFound desc = could not find container \"23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1\": container with ID starting with 23b6bad74850f03775cb14fd6033093dce5e52eef65152c80e2b019e2bbd70b1 not found: ID does not exist" Dec 17 09:32:02 crc kubenswrapper[4966]: I1217 09:32:02.846135 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" path="/var/lib/kubelet/pods/a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a/volumes" Dec 17 09:32:46 crc kubenswrapper[4966]: I1217 09:32:46.808272 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:32:46 crc kubenswrapper[4966]: I1217 09:32:46.809156 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:33:16 crc kubenswrapper[4966]: I1217 09:33:16.807776 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:33:16 crc kubenswrapper[4966]: I1217 09:33:16.808269 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:33:46 crc kubenswrapper[4966]: I1217 09:33:46.807496 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:33:46 crc kubenswrapper[4966]: I1217 09:33:46.808057 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:33:46 crc kubenswrapper[4966]: I1217 09:33:46.808321 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:33:46 crc kubenswrapper[4966]: I1217 09:33:46.809296 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e5578a74a8e5e7a291125ba7509b12de429ea30e5bd8aebceb023dc7cc3a7f9"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:33:46 crc kubenswrapper[4966]: I1217 09:33:46.809722 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://4e5578a74a8e5e7a291125ba7509b12de429ea30e5bd8aebceb023dc7cc3a7f9" gracePeriod=600 Dec 17 09:33:47 crc kubenswrapper[4966]: I1217 09:33:47.091905 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"4e5578a74a8e5e7a291125ba7509b12de429ea30e5bd8aebceb023dc7cc3a7f9"} Dec 17 09:33:47 crc kubenswrapper[4966]: I1217 09:33:47.092182 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="4e5578a74a8e5e7a291125ba7509b12de429ea30e5bd8aebceb023dc7cc3a7f9" exitCode=0 Dec 17 09:33:47 crc kubenswrapper[4966]: I1217 09:33:47.092693 4966 scope.go:117] "RemoveContainer" containerID="c31ba944ea55056ed3acc9f97871d51a6c7e3bc359922d1cca945c0c4d84086e" Dec 17 09:33:48 crc kubenswrapper[4966]: I1217 09:33:48.103117 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e"} Dec 17 09:36:16 crc kubenswrapper[4966]: I1217 09:36:16.807741 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:36:16 crc kubenswrapper[4966]: I1217 09:36:16.808566 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:36:46 crc kubenswrapper[4966]: I1217 09:36:46.807978 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:36:46 crc kubenswrapper[4966]: I1217 09:36:46.808638 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:37:16 crc kubenswrapper[4966]: I1217 09:37:16.807264 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:37:16 crc kubenswrapper[4966]: I1217 09:37:16.807931 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:37:16 crc kubenswrapper[4966]: I1217 09:37:16.807987 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:37:16 crc kubenswrapper[4966]: I1217 09:37:16.808783 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:37:16 crc kubenswrapper[4966]: I1217 09:37:16.808845 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" gracePeriod=600 Dec 17 09:37:16 crc kubenswrapper[4966]: E1217 09:37:16.938929 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:37:17 crc kubenswrapper[4966]: I1217 09:37:17.033039 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" exitCode=0 Dec 17 09:37:17 crc kubenswrapper[4966]: I1217 09:37:17.033242 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e"} Dec 17 09:37:17 crc kubenswrapper[4966]: I1217 09:37:17.033496 4966 scope.go:117] "RemoveContainer" containerID="4e5578a74a8e5e7a291125ba7509b12de429ea30e5bd8aebceb023dc7cc3a7f9" Dec 17 09:37:17 crc kubenswrapper[4966]: I1217 09:37:17.035325 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:37:17 crc kubenswrapper[4966]: E1217 09:37:17.036113 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:37:31 crc kubenswrapper[4966]: I1217 09:37:31.831209 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:37:31 crc kubenswrapper[4966]: E1217 09:37:31.832073 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:37:45 crc kubenswrapper[4966]: I1217 09:37:45.830815 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:37:45 crc kubenswrapper[4966]: E1217 09:37:45.831669 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.366794 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-txhmf"] Dec 17 09:37:55 crc kubenswrapper[4966]: E1217 09:37:55.368931 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerName="extract-content" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.368967 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerName="extract-content" Dec 17 09:37:55 crc kubenswrapper[4966]: E1217 09:37:55.369002 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerName="registry-server" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.369016 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerName="registry-server" Dec 17 09:37:55 crc kubenswrapper[4966]: E1217 09:37:55.369047 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerName="extract-utilities" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.369059 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerName="extract-utilities" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.369446 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2ad4f7e-ecb3-4941-aa7b-70568f0fd15a" containerName="registry-server" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.374330 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.383711 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-catalog-content\") pod \"community-operators-txhmf\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.383760 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqgzw\" (UniqueName: \"kubernetes.io/projected/9a4af493-1dc3-4656-95a1-6cb8174e96c8-kube-api-access-vqgzw\") pod \"community-operators-txhmf\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.383811 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-utilities\") pod \"community-operators-txhmf\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.401108 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-txhmf"] Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.485197 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-catalog-content\") pod \"community-operators-txhmf\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.485241 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqgzw\" (UniqueName: \"kubernetes.io/projected/9a4af493-1dc3-4656-95a1-6cb8174e96c8-kube-api-access-vqgzw\") pod \"community-operators-txhmf\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.485297 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-utilities\") pod \"community-operators-txhmf\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.487406 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-catalog-content\") pod \"community-operators-txhmf\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.487721 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-utilities\") pod \"community-operators-txhmf\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.509384 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqgzw\" (UniqueName: \"kubernetes.io/projected/9a4af493-1dc3-4656-95a1-6cb8174e96c8-kube-api-access-vqgzw\") pod \"community-operators-txhmf\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:55 crc kubenswrapper[4966]: I1217 09:37:55.695142 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:37:56 crc kubenswrapper[4966]: I1217 09:37:56.690725 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-txhmf"] Dec 17 09:37:57 crc kubenswrapper[4966]: I1217 09:37:57.428363 4966 generic.go:334] "Generic (PLEG): container finished" podID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerID="56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17" exitCode=0 Dec 17 09:37:57 crc kubenswrapper[4966]: I1217 09:37:57.428602 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txhmf" event={"ID":"9a4af493-1dc3-4656-95a1-6cb8174e96c8","Type":"ContainerDied","Data":"56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17"} Dec 17 09:37:57 crc kubenswrapper[4966]: I1217 09:37:57.428672 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txhmf" event={"ID":"9a4af493-1dc3-4656-95a1-6cb8174e96c8","Type":"ContainerStarted","Data":"d7149f32e1b51a5dc07d1552be2a46c8f5c235eac6d598952eb85fa9dd901f7e"} Dec 17 09:37:57 crc kubenswrapper[4966]: I1217 09:37:57.430760 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 09:37:57 crc kubenswrapper[4966]: I1217 09:37:57.830332 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:37:57 crc kubenswrapper[4966]: E1217 09:37:57.831836 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:37:58 crc kubenswrapper[4966]: I1217 09:37:58.447380 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txhmf" event={"ID":"9a4af493-1dc3-4656-95a1-6cb8174e96c8","Type":"ContainerStarted","Data":"e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25"} Dec 17 09:37:59 crc kubenswrapper[4966]: I1217 09:37:59.456903 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txhmf" event={"ID":"9a4af493-1dc3-4656-95a1-6cb8174e96c8","Type":"ContainerDied","Data":"e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25"} Dec 17 09:37:59 crc kubenswrapper[4966]: I1217 09:37:59.456864 4966 generic.go:334] "Generic (PLEG): container finished" podID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerID="e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25" exitCode=0 Dec 17 09:38:01 crc kubenswrapper[4966]: I1217 09:38:01.478917 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txhmf" event={"ID":"9a4af493-1dc3-4656-95a1-6cb8174e96c8","Type":"ContainerStarted","Data":"ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc"} Dec 17 09:38:01 crc kubenswrapper[4966]: I1217 09:38:01.504629 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-txhmf" podStartSLOduration=2.787093089 podStartE2EDuration="6.503582996s" podCreationTimestamp="2025-12-17 09:37:55 +0000 UTC" firstStartedPulling="2025-12-17 09:37:57.430036809 +0000 UTC m=+4612.975106761" lastFinishedPulling="2025-12-17 09:38:01.146526706 +0000 UTC m=+4616.691596668" observedRunningTime="2025-12-17 09:38:01.494514079 +0000 UTC m=+4617.039584021" watchObservedRunningTime="2025-12-17 09:38:01.503582996 +0000 UTC m=+4617.048652948" Dec 17 09:38:05 crc kubenswrapper[4966]: I1217 09:38:05.696348 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:38:05 crc kubenswrapper[4966]: I1217 09:38:05.697424 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:38:06 crc kubenswrapper[4966]: I1217 09:38:06.740506 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-txhmf" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerName="registry-server" probeResult="failure" output=< Dec 17 09:38:06 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:38:06 crc kubenswrapper[4966]: > Dec 17 09:38:11 crc kubenswrapper[4966]: I1217 09:38:11.831159 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:38:11 crc kubenswrapper[4966]: E1217 09:38:11.832264 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:38:15 crc kubenswrapper[4966]: I1217 09:38:15.741948 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:38:15 crc kubenswrapper[4966]: I1217 09:38:15.797318 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:38:15 crc kubenswrapper[4966]: I1217 09:38:15.978713 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-txhmf"] Dec 17 09:38:17 crc kubenswrapper[4966]: I1217 09:38:17.664792 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-txhmf" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerName="registry-server" containerID="cri-o://ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc" gracePeriod=2 Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.307506 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.499092 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-utilities\") pod \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.499353 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqgzw\" (UniqueName: \"kubernetes.io/projected/9a4af493-1dc3-4656-95a1-6cb8174e96c8-kube-api-access-vqgzw\") pod \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.499574 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-catalog-content\") pod \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\" (UID: \"9a4af493-1dc3-4656-95a1-6cb8174e96c8\") " Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.500851 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-utilities" (OuterVolumeSpecName: "utilities") pod "9a4af493-1dc3-4656-95a1-6cb8174e96c8" (UID: "9a4af493-1dc3-4656-95a1-6cb8174e96c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.523161 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4af493-1dc3-4656-95a1-6cb8174e96c8-kube-api-access-vqgzw" (OuterVolumeSpecName: "kube-api-access-vqgzw") pod "9a4af493-1dc3-4656-95a1-6cb8174e96c8" (UID: "9a4af493-1dc3-4656-95a1-6cb8174e96c8"). InnerVolumeSpecName "kube-api-access-vqgzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.552473 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a4af493-1dc3-4656-95a1-6cb8174e96c8" (UID: "9a4af493-1dc3-4656-95a1-6cb8174e96c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.601724 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.602224 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a4af493-1dc3-4656-95a1-6cb8174e96c8-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.602342 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqgzw\" (UniqueName: \"kubernetes.io/projected/9a4af493-1dc3-4656-95a1-6cb8174e96c8-kube-api-access-vqgzw\") on node \"crc\" DevicePath \"\"" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.678460 4966 generic.go:334] "Generic (PLEG): container finished" podID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerID="ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc" exitCode=0 Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.678614 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txhmf" event={"ID":"9a4af493-1dc3-4656-95a1-6cb8174e96c8","Type":"ContainerDied","Data":"ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc"} Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.679623 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txhmf" event={"ID":"9a4af493-1dc3-4656-95a1-6cb8174e96c8","Type":"ContainerDied","Data":"d7149f32e1b51a5dc07d1552be2a46c8f5c235eac6d598952eb85fa9dd901f7e"} Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.679797 4966 scope.go:117] "RemoveContainer" containerID="ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.678782 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txhmf" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.742800 4966 scope.go:117] "RemoveContainer" containerID="e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.784762 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-txhmf"] Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.799001 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-txhmf"] Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.803057 4966 scope.go:117] "RemoveContainer" containerID="56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.835840 4966 scope.go:117] "RemoveContainer" containerID="ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc" Dec 17 09:38:18 crc kubenswrapper[4966]: E1217 09:38:18.837949 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc\": container with ID starting with ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc not found: ID does not exist" containerID="ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.838274 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc"} err="failed to get container status \"ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc\": rpc error: code = NotFound desc = could not find container \"ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc\": container with ID starting with ef82a4ec5ad0597d4c80e9b2c847e8db41408b71e3a042f470e570b53dbf05bc not found: ID does not exist" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.838323 4966 scope.go:117] "RemoveContainer" containerID="e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25" Dec 17 09:38:18 crc kubenswrapper[4966]: E1217 09:38:18.838676 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25\": container with ID starting with e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25 not found: ID does not exist" containerID="e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.838708 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25"} err="failed to get container status \"e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25\": rpc error: code = NotFound desc = could not find container \"e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25\": container with ID starting with e3fa678cde63ebbe806473899d3fadc925c756424463af01cbbbc06e0495de25 not found: ID does not exist" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.838728 4966 scope.go:117] "RemoveContainer" containerID="56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17" Dec 17 09:38:18 crc kubenswrapper[4966]: E1217 09:38:18.839134 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17\": container with ID starting with 56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17 not found: ID does not exist" containerID="56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.839164 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17"} err="failed to get container status \"56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17\": rpc error: code = NotFound desc = could not find container \"56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17\": container with ID starting with 56d4050785afc79f3e594fc6ba43a096be44692ed3333beac7b713d974ef6a17 not found: ID does not exist" Dec 17 09:38:18 crc kubenswrapper[4966]: I1217 09:38:18.844146 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" path="/var/lib/kubelet/pods/9a4af493-1dc3-4656-95a1-6cb8174e96c8/volumes" Dec 17 09:38:23 crc kubenswrapper[4966]: I1217 09:38:23.830593 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:38:23 crc kubenswrapper[4966]: E1217 09:38:23.831346 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:38:38 crc kubenswrapper[4966]: I1217 09:38:38.830231 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:38:38 crc kubenswrapper[4966]: E1217 09:38:38.830924 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:38:50 crc kubenswrapper[4966]: I1217 09:38:50.831100 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:38:50 crc kubenswrapper[4966]: E1217 09:38:50.831836 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:39:03 crc kubenswrapper[4966]: I1217 09:39:03.831135 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:39:03 crc kubenswrapper[4966]: E1217 09:39:03.832130 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:39:18 crc kubenswrapper[4966]: I1217 09:39:18.833599 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:39:18 crc kubenswrapper[4966]: E1217 09:39:18.834594 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:39:32 crc kubenswrapper[4966]: I1217 09:39:32.831340 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:39:32 crc kubenswrapper[4966]: E1217 09:39:32.832341 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:39:45 crc kubenswrapper[4966]: I1217 09:39:45.830775 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:39:45 crc kubenswrapper[4966]: E1217 09:39:45.831353 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:39:53 crc kubenswrapper[4966]: I1217 09:39:53.992656 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ctxg7"] Dec 17 09:39:53 crc kubenswrapper[4966]: E1217 09:39:53.994700 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerName="extract-content" Dec 17 09:39:53 crc kubenswrapper[4966]: I1217 09:39:53.994724 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerName="extract-content" Dec 17 09:39:53 crc kubenswrapper[4966]: E1217 09:39:53.994755 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerName="extract-utilities" Dec 17 09:39:53 crc kubenswrapper[4966]: I1217 09:39:53.994762 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerName="extract-utilities" Dec 17 09:39:53 crc kubenswrapper[4966]: E1217 09:39:53.994779 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerName="registry-server" Dec 17 09:39:53 crc kubenswrapper[4966]: I1217 09:39:53.994785 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerName="registry-server" Dec 17 09:39:53 crc kubenswrapper[4966]: I1217 09:39:53.995288 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4af493-1dc3-4656-95a1-6cb8174e96c8" containerName="registry-server" Dec 17 09:39:53 crc kubenswrapper[4966]: I1217 09:39:53.997583 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.016114 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ctxg7"] Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.060199 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-catalog-content\") pod \"redhat-operators-ctxg7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.060556 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-utilities\") pod \"redhat-operators-ctxg7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.060700 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swnn9\" (UniqueName: \"kubernetes.io/projected/bd0f513f-2060-449e-b51f-d3c98fab52f7-kube-api-access-swnn9\") pod \"redhat-operators-ctxg7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.163243 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-catalog-content\") pod \"redhat-operators-ctxg7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.163281 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-utilities\") pod \"redhat-operators-ctxg7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.163334 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swnn9\" (UniqueName: \"kubernetes.io/projected/bd0f513f-2060-449e-b51f-d3c98fab52f7-kube-api-access-swnn9\") pod \"redhat-operators-ctxg7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.165359 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-utilities\") pod \"redhat-operators-ctxg7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.165988 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-catalog-content\") pod \"redhat-operators-ctxg7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.198408 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swnn9\" (UniqueName: \"kubernetes.io/projected/bd0f513f-2060-449e-b51f-d3c98fab52f7-kube-api-access-swnn9\") pod \"redhat-operators-ctxg7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:54 crc kubenswrapper[4966]: I1217 09:39:54.353279 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:39:55 crc kubenswrapper[4966]: I1217 09:39:55.613760 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ctxg7"] Dec 17 09:39:55 crc kubenswrapper[4966]: I1217 09:39:55.687493 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ctxg7" event={"ID":"bd0f513f-2060-449e-b51f-d3c98fab52f7","Type":"ContainerStarted","Data":"9f8dbf3bd0dcfde1b7684d2ec1b319c0a6e8e57d635c1ff3bcd98fbd5dbc8ad2"} Dec 17 09:39:56 crc kubenswrapper[4966]: I1217 09:39:56.712303 4966 generic.go:334] "Generic (PLEG): container finished" podID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerID="6b0d42faef250378aa745d0014f7626b1172ce4bb66dd0a9cba241c4d20f4d5f" exitCode=0 Dec 17 09:39:56 crc kubenswrapper[4966]: I1217 09:39:56.712567 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ctxg7" event={"ID":"bd0f513f-2060-449e-b51f-d3c98fab52f7","Type":"ContainerDied","Data":"6b0d42faef250378aa745d0014f7626b1172ce4bb66dd0a9cba241c4d20f4d5f"} Dec 17 09:39:58 crc kubenswrapper[4966]: I1217 09:39:58.763538 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ctxg7" event={"ID":"bd0f513f-2060-449e-b51f-d3c98fab52f7","Type":"ContainerStarted","Data":"65fa11a1d3aa9a4836a6ec3d45f9c140bd0e12c8df5f42fa9369b7acd3291ff8"} Dec 17 09:39:58 crc kubenswrapper[4966]: I1217 09:39:58.830942 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:39:58 crc kubenswrapper[4966]: E1217 09:39:58.831258 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:40:01 crc kubenswrapper[4966]: I1217 09:40:01.789125 4966 generic.go:334] "Generic (PLEG): container finished" podID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerID="65fa11a1d3aa9a4836a6ec3d45f9c140bd0e12c8df5f42fa9369b7acd3291ff8" exitCode=0 Dec 17 09:40:01 crc kubenswrapper[4966]: I1217 09:40:01.789204 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ctxg7" event={"ID":"bd0f513f-2060-449e-b51f-d3c98fab52f7","Type":"ContainerDied","Data":"65fa11a1d3aa9a4836a6ec3d45f9c140bd0e12c8df5f42fa9369b7acd3291ff8"} Dec 17 09:40:03 crc kubenswrapper[4966]: I1217 09:40:03.810224 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ctxg7" event={"ID":"bd0f513f-2060-449e-b51f-d3c98fab52f7","Type":"ContainerStarted","Data":"b3b3df785d556644a92d207b532ced3c273ddd6114f413104eda3b49106727e0"} Dec 17 09:40:03 crc kubenswrapper[4966]: I1217 09:40:03.839513 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ctxg7" podStartSLOduration=4.696366498 podStartE2EDuration="10.837705052s" podCreationTimestamp="2025-12-17 09:39:53 +0000 UTC" firstStartedPulling="2025-12-17 09:39:56.714969278 +0000 UTC m=+4732.260039220" lastFinishedPulling="2025-12-17 09:40:02.856307842 +0000 UTC m=+4738.401377774" observedRunningTime="2025-12-17 09:40:03.832751767 +0000 UTC m=+4739.377821709" watchObservedRunningTime="2025-12-17 09:40:03.837705052 +0000 UTC m=+4739.382774994" Dec 17 09:40:04 crc kubenswrapper[4966]: I1217 09:40:04.353608 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:40:04 crc kubenswrapper[4966]: I1217 09:40:04.353899 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:40:05 crc kubenswrapper[4966]: I1217 09:40:05.406336 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ctxg7" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="registry-server" probeResult="failure" output=< Dec 17 09:40:05 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:40:05 crc kubenswrapper[4966]: > Dec 17 09:40:13 crc kubenswrapper[4966]: I1217 09:40:13.830481 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:40:13 crc kubenswrapper[4966]: E1217 09:40:13.831575 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:40:15 crc kubenswrapper[4966]: I1217 09:40:15.405319 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ctxg7" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="registry-server" probeResult="failure" output=< Dec 17 09:40:15 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:40:15 crc kubenswrapper[4966]: > Dec 17 09:40:24 crc kubenswrapper[4966]: I1217 09:40:24.408722 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:40:24 crc kubenswrapper[4966]: I1217 09:40:24.481237 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:40:24 crc kubenswrapper[4966]: I1217 09:40:24.840930 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:40:24 crc kubenswrapper[4966]: E1217 09:40:24.855041 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:40:25 crc kubenswrapper[4966]: I1217 09:40:25.189245 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ctxg7"] Dec 17 09:40:26 crc kubenswrapper[4966]: I1217 09:40:26.009635 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ctxg7" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="registry-server" containerID="cri-o://b3b3df785d556644a92d207b532ced3c273ddd6114f413104eda3b49106727e0" gracePeriod=2 Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.032193 4966 generic.go:334] "Generic (PLEG): container finished" podID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerID="b3b3df785d556644a92d207b532ced3c273ddd6114f413104eda3b49106727e0" exitCode=0 Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.032518 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ctxg7" event={"ID":"bd0f513f-2060-449e-b51f-d3c98fab52f7","Type":"ContainerDied","Data":"b3b3df785d556644a92d207b532ced3c273ddd6114f413104eda3b49106727e0"} Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.216369 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.278729 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-utilities\") pod \"bd0f513f-2060-449e-b51f-d3c98fab52f7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.278829 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-catalog-content\") pod \"bd0f513f-2060-449e-b51f-d3c98fab52f7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.278881 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swnn9\" (UniqueName: \"kubernetes.io/projected/bd0f513f-2060-449e-b51f-d3c98fab52f7-kube-api-access-swnn9\") pod \"bd0f513f-2060-449e-b51f-d3c98fab52f7\" (UID: \"bd0f513f-2060-449e-b51f-d3c98fab52f7\") " Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.279632 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-utilities" (OuterVolumeSpecName: "utilities") pod "bd0f513f-2060-449e-b51f-d3c98fab52f7" (UID: "bd0f513f-2060-449e-b51f-d3c98fab52f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.297324 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd0f513f-2060-449e-b51f-d3c98fab52f7-kube-api-access-swnn9" (OuterVolumeSpecName: "kube-api-access-swnn9") pod "bd0f513f-2060-449e-b51f-d3c98fab52f7" (UID: "bd0f513f-2060-449e-b51f-d3c98fab52f7"). InnerVolumeSpecName "kube-api-access-swnn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.380821 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swnn9\" (UniqueName: \"kubernetes.io/projected/bd0f513f-2060-449e-b51f-d3c98fab52f7-kube-api-access-swnn9\") on node \"crc\" DevicePath \"\"" Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.380855 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.418810 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd0f513f-2060-449e-b51f-d3c98fab52f7" (UID: "bd0f513f-2060-449e-b51f-d3c98fab52f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:40:27 crc kubenswrapper[4966]: I1217 09:40:27.482706 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd0f513f-2060-449e-b51f-d3c98fab52f7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:40:28 crc kubenswrapper[4966]: I1217 09:40:28.043949 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ctxg7" event={"ID":"bd0f513f-2060-449e-b51f-d3c98fab52f7","Type":"ContainerDied","Data":"9f8dbf3bd0dcfde1b7684d2ec1b319c0a6e8e57d635c1ff3bcd98fbd5dbc8ad2"} Dec 17 09:40:28 crc kubenswrapper[4966]: I1217 09:40:28.044230 4966 scope.go:117] "RemoveContainer" containerID="b3b3df785d556644a92d207b532ced3c273ddd6114f413104eda3b49106727e0" Dec 17 09:40:28 crc kubenswrapper[4966]: I1217 09:40:28.043997 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ctxg7" Dec 17 09:40:28 crc kubenswrapper[4966]: I1217 09:40:28.065652 4966 scope.go:117] "RemoveContainer" containerID="65fa11a1d3aa9a4836a6ec3d45f9c140bd0e12c8df5f42fa9369b7acd3291ff8" Dec 17 09:40:28 crc kubenswrapper[4966]: I1217 09:40:28.083972 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ctxg7"] Dec 17 09:40:28 crc kubenswrapper[4966]: I1217 09:40:28.093235 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ctxg7"] Dec 17 09:40:28 crc kubenswrapper[4966]: I1217 09:40:28.099332 4966 scope.go:117] "RemoveContainer" containerID="6b0d42faef250378aa745d0014f7626b1172ce4bb66dd0a9cba241c4d20f4d5f" Dec 17 09:40:28 crc kubenswrapper[4966]: I1217 09:40:28.847248 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" path="/var/lib/kubelet/pods/bd0f513f-2060-449e-b51f-d3c98fab52f7/volumes" Dec 17 09:40:35 crc kubenswrapper[4966]: I1217 09:40:35.832499 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:40:35 crc kubenswrapper[4966]: E1217 09:40:35.835533 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:40:48 crc kubenswrapper[4966]: I1217 09:40:48.831402 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:40:48 crc kubenswrapper[4966]: E1217 09:40:48.832623 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:41:03 crc kubenswrapper[4966]: I1217 09:41:03.830410 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:41:03 crc kubenswrapper[4966]: E1217 09:41:03.831198 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:41:16 crc kubenswrapper[4966]: I1217 09:41:16.830215 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:41:16 crc kubenswrapper[4966]: E1217 09:41:16.831075 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.203497 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5pclf"] Dec 17 09:41:28 crc kubenswrapper[4966]: E1217 09:41:28.204363 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="extract-utilities" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.204376 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="extract-utilities" Dec 17 09:41:28 crc kubenswrapper[4966]: E1217 09:41:28.204406 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="extract-content" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.204413 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="extract-content" Dec 17 09:41:28 crc kubenswrapper[4966]: E1217 09:41:28.204430 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="registry-server" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.204437 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="registry-server" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.204603 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd0f513f-2060-449e-b51f-d3c98fab52f7" containerName="registry-server" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.205955 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.240832 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pclf"] Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.366816 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-catalog-content\") pod \"redhat-marketplace-5pclf\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.367394 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-utilities\") pod \"redhat-marketplace-5pclf\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.367649 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsccq\" (UniqueName: \"kubernetes.io/projected/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-kube-api-access-qsccq\") pod \"redhat-marketplace-5pclf\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.469360 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsccq\" (UniqueName: \"kubernetes.io/projected/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-kube-api-access-qsccq\") pod \"redhat-marketplace-5pclf\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.469445 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-catalog-content\") pod \"redhat-marketplace-5pclf\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.469524 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-utilities\") pod \"redhat-marketplace-5pclf\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.470034 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-utilities\") pod \"redhat-marketplace-5pclf\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.470266 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-catalog-content\") pod \"redhat-marketplace-5pclf\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.499345 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsccq\" (UniqueName: \"kubernetes.io/projected/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-kube-api-access-qsccq\") pod \"redhat-marketplace-5pclf\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:28 crc kubenswrapper[4966]: I1217 09:41:28.586482 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:29 crc kubenswrapper[4966]: I1217 09:41:29.107292 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pclf"] Dec 17 09:41:29 crc kubenswrapper[4966]: I1217 09:41:29.669191 4966 generic.go:334] "Generic (PLEG): container finished" podID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerID="0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317" exitCode=0 Dec 17 09:41:29 crc kubenswrapper[4966]: I1217 09:41:29.669431 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pclf" event={"ID":"9cf0d005-00a7-4be6-ab3b-2b440425bcd3","Type":"ContainerDied","Data":"0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317"} Dec 17 09:41:29 crc kubenswrapper[4966]: I1217 09:41:29.669512 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pclf" event={"ID":"9cf0d005-00a7-4be6-ab3b-2b440425bcd3","Type":"ContainerStarted","Data":"a0c9ef193394ebb34793d04f115db18b039ec6c81075b4357603086451c21050"} Dec 17 09:41:30 crc kubenswrapper[4966]: I1217 09:41:30.679685 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pclf" event={"ID":"9cf0d005-00a7-4be6-ab3b-2b440425bcd3","Type":"ContainerStarted","Data":"b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1"} Dec 17 09:41:30 crc kubenswrapper[4966]: I1217 09:41:30.830569 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:41:30 crc kubenswrapper[4966]: E1217 09:41:30.830787 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:41:32 crc kubenswrapper[4966]: I1217 09:41:32.699737 4966 generic.go:334] "Generic (PLEG): container finished" podID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerID="b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1" exitCode=0 Dec 17 09:41:32 crc kubenswrapper[4966]: I1217 09:41:32.699837 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pclf" event={"ID":"9cf0d005-00a7-4be6-ab3b-2b440425bcd3","Type":"ContainerDied","Data":"b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1"} Dec 17 09:41:33 crc kubenswrapper[4966]: I1217 09:41:33.713393 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pclf" event={"ID":"9cf0d005-00a7-4be6-ab3b-2b440425bcd3","Type":"ContainerStarted","Data":"5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c"} Dec 17 09:41:33 crc kubenswrapper[4966]: I1217 09:41:33.743682 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5pclf" podStartSLOduration=2.06503731 podStartE2EDuration="5.743662486s" podCreationTimestamp="2025-12-17 09:41:28 +0000 UTC" firstStartedPulling="2025-12-17 09:41:29.67493057 +0000 UTC m=+4825.220000522" lastFinishedPulling="2025-12-17 09:41:33.353555756 +0000 UTC m=+4828.898625698" observedRunningTime="2025-12-17 09:41:33.743189824 +0000 UTC m=+4829.288259756" watchObservedRunningTime="2025-12-17 09:41:33.743662486 +0000 UTC m=+4829.288732428" Dec 17 09:41:38 crc kubenswrapper[4966]: I1217 09:41:38.586695 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:38 crc kubenswrapper[4966]: I1217 09:41:38.587275 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:38 crc kubenswrapper[4966]: I1217 09:41:38.657014 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:38 crc kubenswrapper[4966]: I1217 09:41:38.816434 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:38 crc kubenswrapper[4966]: I1217 09:41:38.907463 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pclf"] Dec 17 09:41:40 crc kubenswrapper[4966]: I1217 09:41:40.785762 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5pclf" podUID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerName="registry-server" containerID="cri-o://5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c" gracePeriod=2 Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.289909 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.438465 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-utilities\") pod \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.438702 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-catalog-content\") pod \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.438808 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsccq\" (UniqueName: \"kubernetes.io/projected/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-kube-api-access-qsccq\") pod \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\" (UID: \"9cf0d005-00a7-4be6-ab3b-2b440425bcd3\") " Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.439535 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-utilities" (OuterVolumeSpecName: "utilities") pod "9cf0d005-00a7-4be6-ab3b-2b440425bcd3" (UID: "9cf0d005-00a7-4be6-ab3b-2b440425bcd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.447236 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-kube-api-access-qsccq" (OuterVolumeSpecName: "kube-api-access-qsccq") pod "9cf0d005-00a7-4be6-ab3b-2b440425bcd3" (UID: "9cf0d005-00a7-4be6-ab3b-2b440425bcd3"). InnerVolumeSpecName "kube-api-access-qsccq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.481849 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cf0d005-00a7-4be6-ab3b-2b440425bcd3" (UID: "9cf0d005-00a7-4be6-ab3b-2b440425bcd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.541854 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.541925 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.541947 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsccq\" (UniqueName: \"kubernetes.io/projected/9cf0d005-00a7-4be6-ab3b-2b440425bcd3-kube-api-access-qsccq\") on node \"crc\" DevicePath \"\"" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.797580 4966 generic.go:334] "Generic (PLEG): container finished" podID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerID="5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c" exitCode=0 Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.797641 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5pclf" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.797683 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pclf" event={"ID":"9cf0d005-00a7-4be6-ab3b-2b440425bcd3","Type":"ContainerDied","Data":"5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c"} Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.799024 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pclf" event={"ID":"9cf0d005-00a7-4be6-ab3b-2b440425bcd3","Type":"ContainerDied","Data":"a0c9ef193394ebb34793d04f115db18b039ec6c81075b4357603086451c21050"} Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.799063 4966 scope.go:117] "RemoveContainer" containerID="5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.833001 4966 scope.go:117] "RemoveContainer" containerID="b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.843571 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pclf"] Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.862057 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pclf"] Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.867984 4966 scope.go:117] "RemoveContainer" containerID="0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.919080 4966 scope.go:117] "RemoveContainer" containerID="5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c" Dec 17 09:41:41 crc kubenswrapper[4966]: E1217 09:41:41.919522 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c\": container with ID starting with 5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c not found: ID does not exist" containerID="5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.919569 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c"} err="failed to get container status \"5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c\": rpc error: code = NotFound desc = could not find container \"5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c\": container with ID starting with 5cabb840ae6615edb5f9d55df1bdb577eaa7e4d8498c1df62e7f82980039949c not found: ID does not exist" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.919595 4966 scope.go:117] "RemoveContainer" containerID="b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1" Dec 17 09:41:41 crc kubenswrapper[4966]: E1217 09:41:41.919861 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1\": container with ID starting with b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1 not found: ID does not exist" containerID="b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.919899 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1"} err="failed to get container status \"b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1\": rpc error: code = NotFound desc = could not find container \"b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1\": container with ID starting with b45ec17a23138a290a902578ac35ae661dbe0b87927a940f0ecc8d1eafa26ea1 not found: ID does not exist" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.919917 4966 scope.go:117] "RemoveContainer" containerID="0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317" Dec 17 09:41:41 crc kubenswrapper[4966]: E1217 09:41:41.920147 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317\": container with ID starting with 0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317 not found: ID does not exist" containerID="0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317" Dec 17 09:41:41 crc kubenswrapper[4966]: I1217 09:41:41.920175 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317"} err="failed to get container status \"0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317\": rpc error: code = NotFound desc = could not find container \"0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317\": container with ID starting with 0484bc44d805bd0d42f82f613c1426bd722a12d00334943e46c8b5c147bae317 not found: ID does not exist" Dec 17 09:41:42 crc kubenswrapper[4966]: I1217 09:41:42.862325 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" path="/var/lib/kubelet/pods/9cf0d005-00a7-4be6-ab3b-2b440425bcd3/volumes" Dec 17 09:41:43 crc kubenswrapper[4966]: I1217 09:41:43.830559 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:41:43 crc kubenswrapper[4966]: E1217 09:41:43.831195 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:41:58 crc kubenswrapper[4966]: I1217 09:41:58.831466 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:41:58 crc kubenswrapper[4966]: E1217 09:41:58.832511 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:42:09 crc kubenswrapper[4966]: I1217 09:42:09.830920 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:42:09 crc kubenswrapper[4966]: E1217 09:42:09.831817 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:42:23 crc kubenswrapper[4966]: I1217 09:42:23.831282 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:42:24 crc kubenswrapper[4966]: I1217 09:42:24.281388 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"5b71d51b65d64f8d6d10ab0c483fa16ceb246bc9c0baab2a2062446e6f95f1e4"} Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.349047 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qf68x"] Dec 17 09:42:43 crc kubenswrapper[4966]: E1217 09:42:43.349939 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerName="extract-utilities" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.349952 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerName="extract-utilities" Dec 17 09:42:43 crc kubenswrapper[4966]: E1217 09:42:43.349967 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerName="registry-server" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.349973 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerName="registry-server" Dec 17 09:42:43 crc kubenswrapper[4966]: E1217 09:42:43.350007 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerName="extract-content" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.350013 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerName="extract-content" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.350227 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cf0d005-00a7-4be6-ab3b-2b440425bcd3" containerName="registry-server" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.351798 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.400343 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qf68x"] Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.491628 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-utilities\") pod \"certified-operators-qf68x\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.491749 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-catalog-content\") pod \"certified-operators-qf68x\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.491948 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d22zn\" (UniqueName: \"kubernetes.io/projected/4e1029ad-7be0-4958-b185-159515dd01c7-kube-api-access-d22zn\") pod \"certified-operators-qf68x\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.596223 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d22zn\" (UniqueName: \"kubernetes.io/projected/4e1029ad-7be0-4958-b185-159515dd01c7-kube-api-access-d22zn\") pod \"certified-operators-qf68x\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.596319 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-utilities\") pod \"certified-operators-qf68x\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.596408 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-catalog-content\") pod \"certified-operators-qf68x\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.596792 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-utilities\") pod \"certified-operators-qf68x\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.596813 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-catalog-content\") pod \"certified-operators-qf68x\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.617188 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d22zn\" (UniqueName: \"kubernetes.io/projected/4e1029ad-7be0-4958-b185-159515dd01c7-kube-api-access-d22zn\") pod \"certified-operators-qf68x\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:43 crc kubenswrapper[4966]: I1217 09:42:43.706345 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:44 crc kubenswrapper[4966]: I1217 09:42:44.227517 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qf68x"] Dec 17 09:42:44 crc kubenswrapper[4966]: W1217 09:42:44.257993 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e1029ad_7be0_4958_b185_159515dd01c7.slice/crio-c7c377b7a84b13915b4f87f20452b1f4d43381e7e38c428f70f1c3131be25fd4 WatchSource:0}: Error finding container c7c377b7a84b13915b4f87f20452b1f4d43381e7e38c428f70f1c3131be25fd4: Status 404 returned error can't find the container with id c7c377b7a84b13915b4f87f20452b1f4d43381e7e38c428f70f1c3131be25fd4 Dec 17 09:42:44 crc kubenswrapper[4966]: I1217 09:42:44.509154 4966 generic.go:334] "Generic (PLEG): container finished" podID="4e1029ad-7be0-4958-b185-159515dd01c7" containerID="1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c" exitCode=0 Dec 17 09:42:44 crc kubenswrapper[4966]: I1217 09:42:44.509339 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qf68x" event={"ID":"4e1029ad-7be0-4958-b185-159515dd01c7","Type":"ContainerDied","Data":"1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c"} Dec 17 09:42:44 crc kubenswrapper[4966]: I1217 09:42:44.509438 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qf68x" event={"ID":"4e1029ad-7be0-4958-b185-159515dd01c7","Type":"ContainerStarted","Data":"c7c377b7a84b13915b4f87f20452b1f4d43381e7e38c428f70f1c3131be25fd4"} Dec 17 09:42:45 crc kubenswrapper[4966]: I1217 09:42:45.541334 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qf68x" event={"ID":"4e1029ad-7be0-4958-b185-159515dd01c7","Type":"ContainerStarted","Data":"f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149"} Dec 17 09:42:46 crc kubenswrapper[4966]: I1217 09:42:46.559176 4966 generic.go:334] "Generic (PLEG): container finished" podID="4e1029ad-7be0-4958-b185-159515dd01c7" containerID="f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149" exitCode=0 Dec 17 09:42:46 crc kubenswrapper[4966]: I1217 09:42:46.559324 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qf68x" event={"ID":"4e1029ad-7be0-4958-b185-159515dd01c7","Type":"ContainerDied","Data":"f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149"} Dec 17 09:42:47 crc kubenswrapper[4966]: I1217 09:42:47.569386 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qf68x" event={"ID":"4e1029ad-7be0-4958-b185-159515dd01c7","Type":"ContainerStarted","Data":"bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752"} Dec 17 09:42:47 crc kubenswrapper[4966]: I1217 09:42:47.607814 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qf68x" podStartSLOduration=2.017352208 podStartE2EDuration="4.607796151s" podCreationTimestamp="2025-12-17 09:42:43 +0000 UTC" firstStartedPulling="2025-12-17 09:42:44.51111837 +0000 UTC m=+4900.056188312" lastFinishedPulling="2025-12-17 09:42:47.101562323 +0000 UTC m=+4902.646632255" observedRunningTime="2025-12-17 09:42:47.597353777 +0000 UTC m=+4903.142423759" watchObservedRunningTime="2025-12-17 09:42:47.607796151 +0000 UTC m=+4903.152866093" Dec 17 09:42:53 crc kubenswrapper[4966]: I1217 09:42:53.707967 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:53 crc kubenswrapper[4966]: I1217 09:42:53.708714 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:53 crc kubenswrapper[4966]: I1217 09:42:53.780427 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:54 crc kubenswrapper[4966]: I1217 09:42:54.704045 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:54 crc kubenswrapper[4966]: I1217 09:42:54.767172 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qf68x"] Dec 17 09:42:56 crc kubenswrapper[4966]: I1217 09:42:56.655895 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qf68x" podUID="4e1029ad-7be0-4958-b185-159515dd01c7" containerName="registry-server" containerID="cri-o://bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752" gracePeriod=2 Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.192906 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.271156 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-catalog-content\") pod \"4e1029ad-7be0-4958-b185-159515dd01c7\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.271396 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-utilities\") pod \"4e1029ad-7be0-4958-b185-159515dd01c7\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.271466 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d22zn\" (UniqueName: \"kubernetes.io/projected/4e1029ad-7be0-4958-b185-159515dd01c7-kube-api-access-d22zn\") pod \"4e1029ad-7be0-4958-b185-159515dd01c7\" (UID: \"4e1029ad-7be0-4958-b185-159515dd01c7\") " Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.272085 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-utilities" (OuterVolumeSpecName: "utilities") pod "4e1029ad-7be0-4958-b185-159515dd01c7" (UID: "4e1029ad-7be0-4958-b185-159515dd01c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.279324 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e1029ad-7be0-4958-b185-159515dd01c7-kube-api-access-d22zn" (OuterVolumeSpecName: "kube-api-access-d22zn") pod "4e1029ad-7be0-4958-b185-159515dd01c7" (UID: "4e1029ad-7be0-4958-b185-159515dd01c7"). InnerVolumeSpecName "kube-api-access-d22zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.343531 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e1029ad-7be0-4958-b185-159515dd01c7" (UID: "4e1029ad-7be0-4958-b185-159515dd01c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.373774 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.373809 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e1029ad-7be0-4958-b185-159515dd01c7-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.373819 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d22zn\" (UniqueName: \"kubernetes.io/projected/4e1029ad-7be0-4958-b185-159515dd01c7-kube-api-access-d22zn\") on node \"crc\" DevicePath \"\"" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.666819 4966 generic.go:334] "Generic (PLEG): container finished" podID="4e1029ad-7be0-4958-b185-159515dd01c7" containerID="bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752" exitCode=0 Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.666908 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qf68x" event={"ID":"4e1029ad-7be0-4958-b185-159515dd01c7","Type":"ContainerDied","Data":"bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752"} Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.666915 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qf68x" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.666951 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qf68x" event={"ID":"4e1029ad-7be0-4958-b185-159515dd01c7","Type":"ContainerDied","Data":"c7c377b7a84b13915b4f87f20452b1f4d43381e7e38c428f70f1c3131be25fd4"} Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.666975 4966 scope.go:117] "RemoveContainer" containerID="bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.709222 4966 scope.go:117] "RemoveContainer" containerID="f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.709623 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qf68x"] Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.717636 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qf68x"] Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.739780 4966 scope.go:117] "RemoveContainer" containerID="1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.788288 4966 scope.go:117] "RemoveContainer" containerID="bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752" Dec 17 09:42:57 crc kubenswrapper[4966]: E1217 09:42:57.789119 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752\": container with ID starting with bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752 not found: ID does not exist" containerID="bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.789169 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752"} err="failed to get container status \"bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752\": rpc error: code = NotFound desc = could not find container \"bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752\": container with ID starting with bf19b9e715e6a104f75eef07f8862934f5ad73b3f3b4bb2176a268428e803752 not found: ID does not exist" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.789203 4966 scope.go:117] "RemoveContainer" containerID="f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149" Dec 17 09:42:57 crc kubenswrapper[4966]: E1217 09:42:57.799112 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149\": container with ID starting with f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149 not found: ID does not exist" containerID="f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.799169 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149"} err="failed to get container status \"f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149\": rpc error: code = NotFound desc = could not find container \"f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149\": container with ID starting with f7b45b3bf63448ee631b5e13bc5f6f2ad6c14e7b8dc8bbbdbe34e27d52763149 not found: ID does not exist" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.799192 4966 scope.go:117] "RemoveContainer" containerID="1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c" Dec 17 09:42:57 crc kubenswrapper[4966]: E1217 09:42:57.800397 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c\": container with ID starting with 1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c not found: ID does not exist" containerID="1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c" Dec 17 09:42:57 crc kubenswrapper[4966]: I1217 09:42:57.800450 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c"} err="failed to get container status \"1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c\": rpc error: code = NotFound desc = could not find container \"1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c\": container with ID starting with 1f2251f14a66280180ef2e8155daa04c4842a4a7dec01c602e0105bc10842c0c not found: ID does not exist" Dec 17 09:42:58 crc kubenswrapper[4966]: I1217 09:42:58.842820 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e1029ad-7be0-4958-b185-159515dd01c7" path="/var/lib/kubelet/pods/4e1029ad-7be0-4958-b185-159515dd01c7/volumes" Dec 17 09:44:46 crc kubenswrapper[4966]: I1217 09:44:46.808567 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:44:46 crc kubenswrapper[4966]: I1217 09:44:46.813926 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.327442 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6"] Dec 17 09:45:00 crc kubenswrapper[4966]: E1217 09:45:00.332319 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e1029ad-7be0-4958-b185-159515dd01c7" containerName="extract-content" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.332359 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e1029ad-7be0-4958-b185-159515dd01c7" containerName="extract-content" Dec 17 09:45:00 crc kubenswrapper[4966]: E1217 09:45:00.332404 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e1029ad-7be0-4958-b185-159515dd01c7" containerName="registry-server" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.332411 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e1029ad-7be0-4958-b185-159515dd01c7" containerName="registry-server" Dec 17 09:45:00 crc kubenswrapper[4966]: E1217 09:45:00.332421 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e1029ad-7be0-4958-b185-159515dd01c7" containerName="extract-utilities" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.332427 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e1029ad-7be0-4958-b185-159515dd01c7" containerName="extract-utilities" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.332963 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e1029ad-7be0-4958-b185-159515dd01c7" containerName="registry-server" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.334899 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.340652 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.340656 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.389450 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6"] Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.494335 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrpcr\" (UniqueName: \"kubernetes.io/projected/f2670d99-8979-4676-9572-fa1b4112ae40-kube-api-access-lrpcr\") pod \"collect-profiles-29432745-vttc6\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.494384 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2670d99-8979-4676-9572-fa1b4112ae40-secret-volume\") pod \"collect-profiles-29432745-vttc6\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.494406 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2670d99-8979-4676-9572-fa1b4112ae40-config-volume\") pod \"collect-profiles-29432745-vttc6\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.596229 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrpcr\" (UniqueName: \"kubernetes.io/projected/f2670d99-8979-4676-9572-fa1b4112ae40-kube-api-access-lrpcr\") pod \"collect-profiles-29432745-vttc6\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.596275 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2670d99-8979-4676-9572-fa1b4112ae40-secret-volume\") pod \"collect-profiles-29432745-vttc6\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.596295 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2670d99-8979-4676-9572-fa1b4112ae40-config-volume\") pod \"collect-profiles-29432745-vttc6\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.597733 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2670d99-8979-4676-9572-fa1b4112ae40-config-volume\") pod \"collect-profiles-29432745-vttc6\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.606602 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2670d99-8979-4676-9572-fa1b4112ae40-secret-volume\") pod \"collect-profiles-29432745-vttc6\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.624645 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrpcr\" (UniqueName: \"kubernetes.io/projected/f2670d99-8979-4676-9572-fa1b4112ae40-kube-api-access-lrpcr\") pod \"collect-profiles-29432745-vttc6\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:00 crc kubenswrapper[4966]: I1217 09:45:00.657481 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:01 crc kubenswrapper[4966]: I1217 09:45:01.235466 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6"] Dec 17 09:45:02 crc kubenswrapper[4966]: I1217 09:45:02.223812 4966 generic.go:334] "Generic (PLEG): container finished" podID="f2670d99-8979-4676-9572-fa1b4112ae40" containerID="3079373bd0dbc1ab85c12568dc383adaa32011cb34bd9d67a1fc79cb6108d781" exitCode=0 Dec 17 09:45:02 crc kubenswrapper[4966]: I1217 09:45:02.224004 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" event={"ID":"f2670d99-8979-4676-9572-fa1b4112ae40","Type":"ContainerDied","Data":"3079373bd0dbc1ab85c12568dc383adaa32011cb34bd9d67a1fc79cb6108d781"} Dec 17 09:45:02 crc kubenswrapper[4966]: I1217 09:45:02.224537 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" event={"ID":"f2670d99-8979-4676-9572-fa1b4112ae40","Type":"ContainerStarted","Data":"f8ed187f200c4096dbd753952e306b7c47d0a01682c0eda2d4bf540e9bcd2a3e"} Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.638197 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.766817 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2670d99-8979-4676-9572-fa1b4112ae40-config-volume\") pod \"f2670d99-8979-4676-9572-fa1b4112ae40\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.767022 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrpcr\" (UniqueName: \"kubernetes.io/projected/f2670d99-8979-4676-9572-fa1b4112ae40-kube-api-access-lrpcr\") pod \"f2670d99-8979-4676-9572-fa1b4112ae40\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.767072 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2670d99-8979-4676-9572-fa1b4112ae40-secret-volume\") pod \"f2670d99-8979-4676-9572-fa1b4112ae40\" (UID: \"f2670d99-8979-4676-9572-fa1b4112ae40\") " Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.768420 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2670d99-8979-4676-9572-fa1b4112ae40-config-volume" (OuterVolumeSpecName: "config-volume") pod "f2670d99-8979-4676-9572-fa1b4112ae40" (UID: "f2670d99-8979-4676-9572-fa1b4112ae40"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.768964 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2670d99-8979-4676-9572-fa1b4112ae40-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.773431 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2670d99-8979-4676-9572-fa1b4112ae40-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f2670d99-8979-4676-9572-fa1b4112ae40" (UID: "f2670d99-8979-4676-9572-fa1b4112ae40"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.785160 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2670d99-8979-4676-9572-fa1b4112ae40-kube-api-access-lrpcr" (OuterVolumeSpecName: "kube-api-access-lrpcr") pod "f2670d99-8979-4676-9572-fa1b4112ae40" (UID: "f2670d99-8979-4676-9572-fa1b4112ae40"). InnerVolumeSpecName "kube-api-access-lrpcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.871293 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrpcr\" (UniqueName: \"kubernetes.io/projected/f2670d99-8979-4676-9572-fa1b4112ae40-kube-api-access-lrpcr\") on node \"crc\" DevicePath \"\"" Dec 17 09:45:03 crc kubenswrapper[4966]: I1217 09:45:03.871331 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2670d99-8979-4676-9572-fa1b4112ae40-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 09:45:04 crc kubenswrapper[4966]: I1217 09:45:04.247376 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" event={"ID":"f2670d99-8979-4676-9572-fa1b4112ae40","Type":"ContainerDied","Data":"f8ed187f200c4096dbd753952e306b7c47d0a01682c0eda2d4bf540e9bcd2a3e"} Dec 17 09:45:04 crc kubenswrapper[4966]: I1217 09:45:04.247431 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8ed187f200c4096dbd753952e306b7c47d0a01682c0eda2d4bf540e9bcd2a3e" Dec 17 09:45:04 crc kubenswrapper[4966]: I1217 09:45:04.247514 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6" Dec 17 09:45:04 crc kubenswrapper[4966]: I1217 09:45:04.733975 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs"] Dec 17 09:45:04 crc kubenswrapper[4966]: I1217 09:45:04.742316 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432700-f8ffs"] Dec 17 09:45:04 crc kubenswrapper[4966]: I1217 09:45:04.840818 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d0cb6a-efa1-4c3a-bd5e-565685eeea80" path="/var/lib/kubelet/pods/84d0cb6a-efa1-4c3a-bd5e-565685eeea80/volumes" Dec 17 09:45:16 crc kubenswrapper[4966]: I1217 09:45:16.807247 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:45:16 crc kubenswrapper[4966]: I1217 09:45:16.807676 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:45:46 crc kubenswrapper[4966]: I1217 09:45:46.807736 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:45:46 crc kubenswrapper[4966]: I1217 09:45:46.808465 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:45:46 crc kubenswrapper[4966]: I1217 09:45:46.808521 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:45:46 crc kubenswrapper[4966]: I1217 09:45:46.809478 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b71d51b65d64f8d6d10ab0c483fa16ceb246bc9c0baab2a2062446e6f95f1e4"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:45:46 crc kubenswrapper[4966]: I1217 09:45:46.809542 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://5b71d51b65d64f8d6d10ab0c483fa16ceb246bc9c0baab2a2062446e6f95f1e4" gracePeriod=600 Dec 17 09:45:46 crc kubenswrapper[4966]: I1217 09:45:46.957073 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="5b71d51b65d64f8d6d10ab0c483fa16ceb246bc9c0baab2a2062446e6f95f1e4" exitCode=0 Dec 17 09:45:46 crc kubenswrapper[4966]: I1217 09:45:46.957134 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"5b71d51b65d64f8d6d10ab0c483fa16ceb246bc9c0baab2a2062446e6f95f1e4"} Dec 17 09:45:46 crc kubenswrapper[4966]: I1217 09:45:46.957185 4966 scope.go:117] "RemoveContainer" containerID="3243294711f82d86b2d77a8eb8a6c8ea097821032775c3190d95f64c139f7a4e" Dec 17 09:45:47 crc kubenswrapper[4966]: I1217 09:45:47.971371 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8"} Dec 17 09:45:59 crc kubenswrapper[4966]: I1217 09:45:59.745081 4966 scope.go:117] "RemoveContainer" containerID="d7e1c0bd4a32d96280b9f82f2a31ead45b81dc91be7e7afbe8f14844757a3543" Dec 17 09:48:16 crc kubenswrapper[4966]: I1217 09:48:16.807710 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:48:16 crc kubenswrapper[4966]: I1217 09:48:16.808328 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:48:46 crc kubenswrapper[4966]: I1217 09:48:46.807532 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:48:46 crc kubenswrapper[4966]: I1217 09:48:46.808184 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:49:16 crc kubenswrapper[4966]: I1217 09:49:16.807916 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:49:16 crc kubenswrapper[4966]: I1217 09:49:16.809325 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:49:16 crc kubenswrapper[4966]: I1217 09:49:16.809465 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:49:16 crc kubenswrapper[4966]: I1217 09:49:16.810284 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:49:16 crc kubenswrapper[4966]: I1217 09:49:16.810622 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" gracePeriod=600 Dec 17 09:49:16 crc kubenswrapper[4966]: E1217 09:49:16.939430 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:49:16 crc kubenswrapper[4966]: I1217 09:49:16.952294 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" exitCode=0 Dec 17 09:49:16 crc kubenswrapper[4966]: I1217 09:49:16.952354 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8"} Dec 17 09:49:16 crc kubenswrapper[4966]: I1217 09:49:16.952464 4966 scope.go:117] "RemoveContainer" containerID="5b71d51b65d64f8d6d10ab0c483fa16ceb246bc9c0baab2a2062446e6f95f1e4" Dec 17 09:49:16 crc kubenswrapper[4966]: I1217 09:49:16.953305 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:49:16 crc kubenswrapper[4966]: E1217 09:49:16.953632 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:49:30 crc kubenswrapper[4966]: I1217 09:49:30.831287 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:49:30 crc kubenswrapper[4966]: E1217 09:49:30.832286 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:49:44 crc kubenswrapper[4966]: I1217 09:49:44.843862 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:49:44 crc kubenswrapper[4966]: E1217 09:49:44.844613 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:49:57 crc kubenswrapper[4966]: I1217 09:49:57.831781 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:49:57 crc kubenswrapper[4966]: E1217 09:49:57.832662 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:50:12 crc kubenswrapper[4966]: I1217 09:50:12.831477 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:50:12 crc kubenswrapper[4966]: E1217 09:50:12.832228 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:50:27 crc kubenswrapper[4966]: I1217 09:50:27.831727 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:50:27 crc kubenswrapper[4966]: E1217 09:50:27.832569 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:50:33 crc kubenswrapper[4966]: I1217 09:50:33.849909 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tx8js"] Dec 17 09:50:33 crc kubenswrapper[4966]: E1217 09:50:33.850974 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2670d99-8979-4676-9572-fa1b4112ae40" containerName="collect-profiles" Dec 17 09:50:33 crc kubenswrapper[4966]: I1217 09:50:33.850989 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2670d99-8979-4676-9572-fa1b4112ae40" containerName="collect-profiles" Dec 17 09:50:33 crc kubenswrapper[4966]: I1217 09:50:33.851244 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2670d99-8979-4676-9572-fa1b4112ae40" containerName="collect-profiles" Dec 17 09:50:33 crc kubenswrapper[4966]: I1217 09:50:33.854167 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:33 crc kubenswrapper[4966]: I1217 09:50:33.861301 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tx8js"] Dec 17 09:50:33 crc kubenswrapper[4966]: I1217 09:50:33.987720 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-catalog-content\") pod \"redhat-operators-tx8js\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:33 crc kubenswrapper[4966]: I1217 09:50:33.987842 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-utilities\") pod \"redhat-operators-tx8js\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:33 crc kubenswrapper[4966]: I1217 09:50:33.988101 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8srnz\" (UniqueName: \"kubernetes.io/projected/15d75032-3db4-46cb-b199-69f7feb92289-kube-api-access-8srnz\") pod \"redhat-operators-tx8js\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:34 crc kubenswrapper[4966]: I1217 09:50:34.089855 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-utilities\") pod \"redhat-operators-tx8js\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:34 crc kubenswrapper[4966]: I1217 09:50:34.090053 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8srnz\" (UniqueName: \"kubernetes.io/projected/15d75032-3db4-46cb-b199-69f7feb92289-kube-api-access-8srnz\") pod \"redhat-operators-tx8js\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:34 crc kubenswrapper[4966]: I1217 09:50:34.090428 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-utilities\") pod \"redhat-operators-tx8js\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:34 crc kubenswrapper[4966]: I1217 09:50:34.090513 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-catalog-content\") pod \"redhat-operators-tx8js\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:34 crc kubenswrapper[4966]: I1217 09:50:34.090854 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-catalog-content\") pod \"redhat-operators-tx8js\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:34 crc kubenswrapper[4966]: I1217 09:50:34.116705 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8srnz\" (UniqueName: \"kubernetes.io/projected/15d75032-3db4-46cb-b199-69f7feb92289-kube-api-access-8srnz\") pod \"redhat-operators-tx8js\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:34 crc kubenswrapper[4966]: I1217 09:50:34.174543 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:34 crc kubenswrapper[4966]: I1217 09:50:34.758754 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tx8js"] Dec 17 09:50:34 crc kubenswrapper[4966]: W1217 09:50:34.766768 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15d75032_3db4_46cb_b199_69f7feb92289.slice/crio-ae8bb2913b5ffbd40f41656240ddfede772ab71b414cfb8f11941572d661498f WatchSource:0}: Error finding container ae8bb2913b5ffbd40f41656240ddfede772ab71b414cfb8f11941572d661498f: Status 404 returned error can't find the container with id ae8bb2913b5ffbd40f41656240ddfede772ab71b414cfb8f11941572d661498f Dec 17 09:50:35 crc kubenswrapper[4966]: I1217 09:50:35.759681 4966 generic.go:334] "Generic (PLEG): container finished" podID="15d75032-3db4-46cb-b199-69f7feb92289" containerID="91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93" exitCode=0 Dec 17 09:50:35 crc kubenswrapper[4966]: I1217 09:50:35.759732 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tx8js" event={"ID":"15d75032-3db4-46cb-b199-69f7feb92289","Type":"ContainerDied","Data":"91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93"} Dec 17 09:50:35 crc kubenswrapper[4966]: I1217 09:50:35.759764 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tx8js" event={"ID":"15d75032-3db4-46cb-b199-69f7feb92289","Type":"ContainerStarted","Data":"ae8bb2913b5ffbd40f41656240ddfede772ab71b414cfb8f11941572d661498f"} Dec 17 09:50:35 crc kubenswrapper[4966]: I1217 09:50:35.763971 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 09:50:36 crc kubenswrapper[4966]: I1217 09:50:36.770163 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tx8js" event={"ID":"15d75032-3db4-46cb-b199-69f7feb92289","Type":"ContainerStarted","Data":"b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f"} Dec 17 09:50:40 crc kubenswrapper[4966]: I1217 09:50:40.811164 4966 generic.go:334] "Generic (PLEG): container finished" podID="15d75032-3db4-46cb-b199-69f7feb92289" containerID="b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f" exitCode=0 Dec 17 09:50:40 crc kubenswrapper[4966]: I1217 09:50:40.811230 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tx8js" event={"ID":"15d75032-3db4-46cb-b199-69f7feb92289","Type":"ContainerDied","Data":"b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f"} Dec 17 09:50:41 crc kubenswrapper[4966]: I1217 09:50:41.821659 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tx8js" event={"ID":"15d75032-3db4-46cb-b199-69f7feb92289","Type":"ContainerStarted","Data":"a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a"} Dec 17 09:50:41 crc kubenswrapper[4966]: I1217 09:50:41.830940 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:50:41 crc kubenswrapper[4966]: E1217 09:50:41.831181 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:50:41 crc kubenswrapper[4966]: I1217 09:50:41.849674 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tx8js" podStartSLOduration=3.329524363 podStartE2EDuration="8.849070955s" podCreationTimestamp="2025-12-17 09:50:33 +0000 UTC" firstStartedPulling="2025-12-17 09:50:35.762318071 +0000 UTC m=+5371.307388013" lastFinishedPulling="2025-12-17 09:50:41.281864663 +0000 UTC m=+5376.826934605" observedRunningTime="2025-12-17 09:50:41.848368275 +0000 UTC m=+5377.393438217" watchObservedRunningTime="2025-12-17 09:50:41.849070955 +0000 UTC m=+5377.394140887" Dec 17 09:50:44 crc kubenswrapper[4966]: I1217 09:50:44.174937 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:44 crc kubenswrapper[4966]: I1217 09:50:44.176401 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:45 crc kubenswrapper[4966]: I1217 09:50:45.224396 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tx8js" podUID="15d75032-3db4-46cb-b199-69f7feb92289" containerName="registry-server" probeResult="failure" output=< Dec 17 09:50:45 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:50:45 crc kubenswrapper[4966]: > Dec 17 09:50:54 crc kubenswrapper[4966]: I1217 09:50:54.236210 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:54 crc kubenswrapper[4966]: I1217 09:50:54.284161 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:54 crc kubenswrapper[4966]: I1217 09:50:54.482284 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tx8js"] Dec 17 09:50:54 crc kubenswrapper[4966]: I1217 09:50:54.839174 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:50:54 crc kubenswrapper[4966]: E1217 09:50:54.839569 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:50:55 crc kubenswrapper[4966]: I1217 09:50:55.951220 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tx8js" podUID="15d75032-3db4-46cb-b199-69f7feb92289" containerName="registry-server" containerID="cri-o://a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a" gracePeriod=2 Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.562698 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.701197 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-catalog-content\") pod \"15d75032-3db4-46cb-b199-69f7feb92289\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.701762 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8srnz\" (UniqueName: \"kubernetes.io/projected/15d75032-3db4-46cb-b199-69f7feb92289-kube-api-access-8srnz\") pod \"15d75032-3db4-46cb-b199-69f7feb92289\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.702024 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-utilities\") pod \"15d75032-3db4-46cb-b199-69f7feb92289\" (UID: \"15d75032-3db4-46cb-b199-69f7feb92289\") " Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.703489 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-utilities" (OuterVolumeSpecName: "utilities") pod "15d75032-3db4-46cb-b199-69f7feb92289" (UID: "15d75032-3db4-46cb-b199-69f7feb92289"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.710293 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15d75032-3db4-46cb-b199-69f7feb92289-kube-api-access-8srnz" (OuterVolumeSpecName: "kube-api-access-8srnz") pod "15d75032-3db4-46cb-b199-69f7feb92289" (UID: "15d75032-3db4-46cb-b199-69f7feb92289"). InnerVolumeSpecName "kube-api-access-8srnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.805472 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.805509 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8srnz\" (UniqueName: \"kubernetes.io/projected/15d75032-3db4-46cb-b199-69f7feb92289-kube-api-access-8srnz\") on node \"crc\" DevicePath \"\"" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.824904 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15d75032-3db4-46cb-b199-69f7feb92289" (UID: "15d75032-3db4-46cb-b199-69f7feb92289"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.907146 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d75032-3db4-46cb-b199-69f7feb92289-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.960594 4966 generic.go:334] "Generic (PLEG): container finished" podID="15d75032-3db4-46cb-b199-69f7feb92289" containerID="a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a" exitCode=0 Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.960637 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tx8js" event={"ID":"15d75032-3db4-46cb-b199-69f7feb92289","Type":"ContainerDied","Data":"a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a"} Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.960650 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tx8js" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.960670 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tx8js" event={"ID":"15d75032-3db4-46cb-b199-69f7feb92289","Type":"ContainerDied","Data":"ae8bb2913b5ffbd40f41656240ddfede772ab71b414cfb8f11941572d661498f"} Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.960690 4966 scope.go:117] "RemoveContainer" containerID="a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.994034 4966 scope.go:117] "RemoveContainer" containerID="b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f" Dec 17 09:50:56 crc kubenswrapper[4966]: I1217 09:50:56.995211 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tx8js"] Dec 17 09:50:57 crc kubenswrapper[4966]: I1217 09:50:57.005916 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tx8js"] Dec 17 09:50:57 crc kubenswrapper[4966]: I1217 09:50:57.049031 4966 scope.go:117] "RemoveContainer" containerID="91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93" Dec 17 09:50:57 crc kubenswrapper[4966]: I1217 09:50:57.092367 4966 scope.go:117] "RemoveContainer" containerID="a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a" Dec 17 09:50:57 crc kubenswrapper[4966]: E1217 09:50:57.093389 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a\": container with ID starting with a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a not found: ID does not exist" containerID="a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a" Dec 17 09:50:57 crc kubenswrapper[4966]: I1217 09:50:57.093442 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a"} err="failed to get container status \"a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a\": rpc error: code = NotFound desc = could not find container \"a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a\": container with ID starting with a224d58d222d85ef0279b0e22fb612a4c17c09c7159b7fafacd438f74e583d3a not found: ID does not exist" Dec 17 09:50:57 crc kubenswrapper[4966]: I1217 09:50:57.093470 4966 scope.go:117] "RemoveContainer" containerID="b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f" Dec 17 09:50:57 crc kubenswrapper[4966]: E1217 09:50:57.093826 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f\": container with ID starting with b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f not found: ID does not exist" containerID="b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f" Dec 17 09:50:57 crc kubenswrapper[4966]: I1217 09:50:57.093881 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f"} err="failed to get container status \"b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f\": rpc error: code = NotFound desc = could not find container \"b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f\": container with ID starting with b9e6ac21e6ce731a211e8b3c8153adb5af543f3614c1a5c81b29b1f80cd5a36f not found: ID does not exist" Dec 17 09:50:57 crc kubenswrapper[4966]: I1217 09:50:57.093903 4966 scope.go:117] "RemoveContainer" containerID="91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93" Dec 17 09:50:57 crc kubenswrapper[4966]: E1217 09:50:57.094190 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93\": container with ID starting with 91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93 not found: ID does not exist" containerID="91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93" Dec 17 09:50:57 crc kubenswrapper[4966]: I1217 09:50:57.094230 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93"} err="failed to get container status \"91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93\": rpc error: code = NotFound desc = could not find container \"91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93\": container with ID starting with 91e302a18f03c8930e2c521f3c5baf627a2541104a3a3ccd9bf7095f15e70f93 not found: ID does not exist" Dec 17 09:50:58 crc kubenswrapper[4966]: I1217 09:50:58.840610 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15d75032-3db4-46cb-b199-69f7feb92289" path="/var/lib/kubelet/pods/15d75032-3db4-46cb-b199-69f7feb92289/volumes" Dec 17 09:51:09 crc kubenswrapper[4966]: I1217 09:51:09.830682 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:51:09 crc kubenswrapper[4966]: E1217 09:51:09.831558 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:51:21 crc kubenswrapper[4966]: I1217 09:51:21.838277 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:51:21 crc kubenswrapper[4966]: E1217 09:51:21.839265 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:51:32 crc kubenswrapper[4966]: I1217 09:51:32.830462 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:51:32 crc kubenswrapper[4966]: E1217 09:51:32.831236 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:51:45 crc kubenswrapper[4966]: I1217 09:51:45.830538 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:51:45 crc kubenswrapper[4966]: E1217 09:51:45.831369 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:52:00 crc kubenswrapper[4966]: I1217 09:52:00.837444 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:52:00 crc kubenswrapper[4966]: E1217 09:52:00.838215 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:52:13 crc kubenswrapper[4966]: I1217 09:52:13.831796 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:52:13 crc kubenswrapper[4966]: E1217 09:52:13.832690 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:52:24 crc kubenswrapper[4966]: I1217 09:52:24.836849 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:52:24 crc kubenswrapper[4966]: E1217 09:52:24.837671 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:52:39 crc kubenswrapper[4966]: I1217 09:52:39.830509 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:52:39 crc kubenswrapper[4966]: E1217 09:52:39.832312 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:52:50 crc kubenswrapper[4966]: I1217 09:52:50.842572 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:52:50 crc kubenswrapper[4966]: E1217 09:52:50.843334 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.038197 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2dsk4"] Dec 17 09:52:52 crc kubenswrapper[4966]: E1217 09:52:52.038698 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d75032-3db4-46cb-b199-69f7feb92289" containerName="registry-server" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.038714 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d75032-3db4-46cb-b199-69f7feb92289" containerName="registry-server" Dec 17 09:52:52 crc kubenswrapper[4966]: E1217 09:52:52.038748 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d75032-3db4-46cb-b199-69f7feb92289" containerName="extract-content" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.038754 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d75032-3db4-46cb-b199-69f7feb92289" containerName="extract-content" Dec 17 09:52:52 crc kubenswrapper[4966]: E1217 09:52:52.038767 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d75032-3db4-46cb-b199-69f7feb92289" containerName="extract-utilities" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.038775 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d75032-3db4-46cb-b199-69f7feb92289" containerName="extract-utilities" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.038987 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="15d75032-3db4-46cb-b199-69f7feb92289" containerName="registry-server" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.040391 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.042563 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-catalog-content\") pod \"redhat-marketplace-2dsk4\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.042736 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvwn2\" (UniqueName: \"kubernetes.io/projected/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-kube-api-access-wvwn2\") pod \"redhat-marketplace-2dsk4\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.042782 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-utilities\") pod \"redhat-marketplace-2dsk4\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.063844 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dsk4"] Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.145224 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-catalog-content\") pod \"redhat-marketplace-2dsk4\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.145426 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvwn2\" (UniqueName: \"kubernetes.io/projected/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-kube-api-access-wvwn2\") pod \"redhat-marketplace-2dsk4\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.145474 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-utilities\") pod \"redhat-marketplace-2dsk4\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.145762 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-catalog-content\") pod \"redhat-marketplace-2dsk4\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.147549 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-utilities\") pod \"redhat-marketplace-2dsk4\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.178213 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvwn2\" (UniqueName: \"kubernetes.io/projected/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-kube-api-access-wvwn2\") pod \"redhat-marketplace-2dsk4\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.362399 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:52:52 crc kubenswrapper[4966]: I1217 09:52:52.936350 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dsk4"] Dec 17 09:52:53 crc kubenswrapper[4966]: I1217 09:52:53.157094 4966 generic.go:334] "Generic (PLEG): container finished" podID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerID="13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66" exitCode=0 Dec 17 09:52:53 crc kubenswrapper[4966]: I1217 09:52:53.157140 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dsk4" event={"ID":"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c","Type":"ContainerDied","Data":"13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66"} Dec 17 09:52:53 crc kubenswrapper[4966]: I1217 09:52:53.157164 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dsk4" event={"ID":"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c","Type":"ContainerStarted","Data":"ca487238df7edc9c4069fc62cfe269ecf3c0c06816be2b24ad6a169227c0127d"} Dec 17 09:52:54 crc kubenswrapper[4966]: I1217 09:52:54.169074 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dsk4" event={"ID":"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c","Type":"ContainerStarted","Data":"c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce"} Dec 17 09:52:55 crc kubenswrapper[4966]: I1217 09:52:55.187418 4966 generic.go:334] "Generic (PLEG): container finished" podID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerID="c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce" exitCode=0 Dec 17 09:52:55 crc kubenswrapper[4966]: I1217 09:52:55.187509 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dsk4" event={"ID":"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c","Type":"ContainerDied","Data":"c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce"} Dec 17 09:52:56 crc kubenswrapper[4966]: I1217 09:52:56.202475 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dsk4" event={"ID":"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c","Type":"ContainerStarted","Data":"2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6"} Dec 17 09:52:56 crc kubenswrapper[4966]: I1217 09:52:56.223132 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2dsk4" podStartSLOduration=1.747966115 podStartE2EDuration="4.223108102s" podCreationTimestamp="2025-12-17 09:52:52 +0000 UTC" firstStartedPulling="2025-12-17 09:52:53.158845395 +0000 UTC m=+5508.703915337" lastFinishedPulling="2025-12-17 09:52:55.633987382 +0000 UTC m=+5511.179057324" observedRunningTime="2025-12-17 09:52:56.219120344 +0000 UTC m=+5511.764190286" watchObservedRunningTime="2025-12-17 09:52:56.223108102 +0000 UTC m=+5511.768178054" Dec 17 09:53:02 crc kubenswrapper[4966]: I1217 09:53:02.362612 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:53:02 crc kubenswrapper[4966]: I1217 09:53:02.363118 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:53:02 crc kubenswrapper[4966]: I1217 09:53:02.419823 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:53:03 crc kubenswrapper[4966]: I1217 09:53:03.343689 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:53:03 crc kubenswrapper[4966]: I1217 09:53:03.402648 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dsk4"] Dec 17 09:53:04 crc kubenswrapper[4966]: I1217 09:53:04.837506 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:53:04 crc kubenswrapper[4966]: E1217 09:53:04.837836 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.312241 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2dsk4" podUID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerName="registry-server" containerID="cri-o://2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6" gracePeriod=2 Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.844621 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.852618 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-utilities\") pod \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.852704 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvwn2\" (UniqueName: \"kubernetes.io/projected/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-kube-api-access-wvwn2\") pod \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.852802 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-catalog-content\") pod \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\" (UID: \"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c\") " Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.854435 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-utilities" (OuterVolumeSpecName: "utilities") pod "2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" (UID: "2abd3c1d-23c1-41a6-a1fa-7cdd1127600c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.861378 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-kube-api-access-wvwn2" (OuterVolumeSpecName: "kube-api-access-wvwn2") pod "2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" (UID: "2abd3c1d-23c1-41a6-a1fa-7cdd1127600c"). InnerVolumeSpecName "kube-api-access-wvwn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.880247 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" (UID: "2abd3c1d-23c1-41a6-a1fa-7cdd1127600c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.954694 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.954741 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvwn2\" (UniqueName: \"kubernetes.io/projected/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-kube-api-access-wvwn2\") on node \"crc\" DevicePath \"\"" Dec 17 09:53:05 crc kubenswrapper[4966]: I1217 09:53:05.954755 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.321756 4966 generic.go:334] "Generic (PLEG): container finished" podID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerID="2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6" exitCode=0 Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.321793 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dsk4" event={"ID":"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c","Type":"ContainerDied","Data":"2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6"} Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.321817 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dsk4" event={"ID":"2abd3c1d-23c1-41a6-a1fa-7cdd1127600c","Type":"ContainerDied","Data":"ca487238df7edc9c4069fc62cfe269ecf3c0c06816be2b24ad6a169227c0127d"} Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.321832 4966 scope.go:117] "RemoveContainer" containerID="2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.321982 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dsk4" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.363782 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dsk4"] Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.370964 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dsk4"] Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.373143 4966 scope.go:117] "RemoveContainer" containerID="c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.408804 4966 scope.go:117] "RemoveContainer" containerID="13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.467675 4966 scope.go:117] "RemoveContainer" containerID="2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6" Dec 17 09:53:06 crc kubenswrapper[4966]: E1217 09:53:06.468483 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6\": container with ID starting with 2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6 not found: ID does not exist" containerID="2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.468534 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6"} err="failed to get container status \"2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6\": rpc error: code = NotFound desc = could not find container \"2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6\": container with ID starting with 2e99f3b4181ecf5ef8c25e5a53e4d541b4343747d27e31f539ba11a578de83b6 not found: ID does not exist" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.468566 4966 scope.go:117] "RemoveContainer" containerID="c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce" Dec 17 09:53:06 crc kubenswrapper[4966]: E1217 09:53:06.468938 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce\": container with ID starting with c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce not found: ID does not exist" containerID="c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.468960 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce"} err="failed to get container status \"c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce\": rpc error: code = NotFound desc = could not find container \"c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce\": container with ID starting with c29b0cb64434ce2c445736ed27993698fa4cc0da9d4b1e215fc5dc86205819ce not found: ID does not exist" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.468974 4966 scope.go:117] "RemoveContainer" containerID="13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66" Dec 17 09:53:06 crc kubenswrapper[4966]: E1217 09:53:06.469382 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66\": container with ID starting with 13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66 not found: ID does not exist" containerID="13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.469428 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66"} err="failed to get container status \"13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66\": rpc error: code = NotFound desc = could not find container \"13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66\": container with ID starting with 13b703e634c1041c727bd109ce19bd394e8da02f3975e7d37e87053a66978f66 not found: ID does not exist" Dec 17 09:53:06 crc kubenswrapper[4966]: I1217 09:53:06.843117 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" path="/var/lib/kubelet/pods/2abd3c1d-23c1-41a6-a1fa-7cdd1127600c/volumes" Dec 17 09:53:19 crc kubenswrapper[4966]: I1217 09:53:19.831715 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:53:19 crc kubenswrapper[4966]: E1217 09:53:19.832576 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:53:30 crc kubenswrapper[4966]: I1217 09:53:30.830788 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:53:30 crc kubenswrapper[4966]: E1217 09:53:30.832183 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:53:45 crc kubenswrapper[4966]: I1217 09:53:45.830435 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:53:45 crc kubenswrapper[4966]: E1217 09:53:45.831215 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:53:58 crc kubenswrapper[4966]: I1217 09:53:58.830350 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:53:58 crc kubenswrapper[4966]: E1217 09:53:58.831074 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:54:11 crc kubenswrapper[4966]: I1217 09:54:11.830653 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:54:11 crc kubenswrapper[4966]: E1217 09:54:11.831319 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 09:54:12 crc kubenswrapper[4966]: I1217 09:54:12.649337 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hzjvh"] Dec 17 09:54:12 crc kubenswrapper[4966]: E1217 09:54:12.686139 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerName="extract-content" Dec 17 09:54:12 crc kubenswrapper[4966]: I1217 09:54:12.686181 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerName="extract-content" Dec 17 09:54:12 crc kubenswrapper[4966]: E1217 09:54:12.686226 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerName="extract-utilities" Dec 17 09:54:12 crc kubenswrapper[4966]: I1217 09:54:12.686239 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerName="extract-utilities" Dec 17 09:54:12 crc kubenswrapper[4966]: E1217 09:54:12.686260 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerName="registry-server" Dec 17 09:54:12 crc kubenswrapper[4966]: I1217 09:54:12.686270 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerName="registry-server" Dec 17 09:54:12 crc kubenswrapper[4966]: I1217 09:54:12.688600 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="2abd3c1d-23c1-41a6-a1fa-7cdd1127600c" containerName="registry-server" Dec 17 09:54:12 crc kubenswrapper[4966]: I1217 09:54:12.867287 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hzjvh"] Dec 17 09:54:12 crc kubenswrapper[4966]: I1217 09:54:12.869220 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.057691 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsb8v\" (UniqueName: \"kubernetes.io/projected/dc5f2752-1355-4f74-9566-b6ecb5a98b96-kube-api-access-dsb8v\") pod \"certified-operators-hzjvh\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.057770 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-catalog-content\") pod \"certified-operators-hzjvh\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.058102 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-utilities\") pod \"certified-operators-hzjvh\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.160838 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-utilities\") pod \"certified-operators-hzjvh\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.161282 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsb8v\" (UniqueName: \"kubernetes.io/projected/dc5f2752-1355-4f74-9566-b6ecb5a98b96-kube-api-access-dsb8v\") pod \"certified-operators-hzjvh\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.161339 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-catalog-content\") pod \"certified-operators-hzjvh\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.161664 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-utilities\") pod \"certified-operators-hzjvh\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.163758 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-catalog-content\") pod \"certified-operators-hzjvh\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.188671 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsb8v\" (UniqueName: \"kubernetes.io/projected/dc5f2752-1355-4f74-9566-b6ecb5a98b96-kube-api-access-dsb8v\") pod \"certified-operators-hzjvh\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.194309 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.666890 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hzjvh"] Dec 17 09:54:13 crc kubenswrapper[4966]: I1217 09:54:13.972138 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzjvh" event={"ID":"dc5f2752-1355-4f74-9566-b6ecb5a98b96","Type":"ContainerStarted","Data":"ed8e21da9385fbcd6f1c9a72b7d6aac83af0dffc8a01705c8b2babcc19e0867a"} Dec 17 09:54:14 crc kubenswrapper[4966]: I1217 09:54:14.982123 4966 generic.go:334] "Generic (PLEG): container finished" podID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerID="237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b" exitCode=0 Dec 17 09:54:14 crc kubenswrapper[4966]: I1217 09:54:14.982216 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzjvh" event={"ID":"dc5f2752-1355-4f74-9566-b6ecb5a98b96","Type":"ContainerDied","Data":"237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b"} Dec 17 09:54:17 crc kubenswrapper[4966]: I1217 09:54:17.026132 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzjvh" event={"ID":"dc5f2752-1355-4f74-9566-b6ecb5a98b96","Type":"ContainerStarted","Data":"a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467"} Dec 17 09:54:18 crc kubenswrapper[4966]: I1217 09:54:18.036176 4966 generic.go:334] "Generic (PLEG): container finished" podID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerID="a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467" exitCode=0 Dec 17 09:54:18 crc kubenswrapper[4966]: I1217 09:54:18.036249 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzjvh" event={"ID":"dc5f2752-1355-4f74-9566-b6ecb5a98b96","Type":"ContainerDied","Data":"a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467"} Dec 17 09:54:19 crc kubenswrapper[4966]: I1217 09:54:19.047143 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzjvh" event={"ID":"dc5f2752-1355-4f74-9566-b6ecb5a98b96","Type":"ContainerStarted","Data":"29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc"} Dec 17 09:54:19 crc kubenswrapper[4966]: I1217 09:54:19.074649 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hzjvh" podStartSLOduration=3.502831275 podStartE2EDuration="7.074623566s" podCreationTimestamp="2025-12-17 09:54:12 +0000 UTC" firstStartedPulling="2025-12-17 09:54:14.984539847 +0000 UTC m=+5590.529609779" lastFinishedPulling="2025-12-17 09:54:18.556332128 +0000 UTC m=+5594.101402070" observedRunningTime="2025-12-17 09:54:19.068324164 +0000 UTC m=+5594.613394116" watchObservedRunningTime="2025-12-17 09:54:19.074623566 +0000 UTC m=+5594.619693508" Dec 17 09:54:23 crc kubenswrapper[4966]: I1217 09:54:23.194511 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:23 crc kubenswrapper[4966]: I1217 09:54:23.195163 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:23 crc kubenswrapper[4966]: I1217 09:54:23.239209 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:23 crc kubenswrapper[4966]: I1217 09:54:23.831437 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:54:24 crc kubenswrapper[4966]: I1217 09:54:24.150059 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:24 crc kubenswrapper[4966]: I1217 09:54:24.204320 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hzjvh"] Dec 17 09:54:25 crc kubenswrapper[4966]: I1217 09:54:25.108035 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"eccaf98bd2744c8937c39460a50fe0df9f527a236e1a3c4c46bd605805659ca1"} Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.116306 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hzjvh" podUID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerName="registry-server" containerID="cri-o://29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc" gracePeriod=2 Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.667690 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.836827 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-utilities\") pod \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.837014 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-catalog-content\") pod \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.837040 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsb8v\" (UniqueName: \"kubernetes.io/projected/dc5f2752-1355-4f74-9566-b6ecb5a98b96-kube-api-access-dsb8v\") pod \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\" (UID: \"dc5f2752-1355-4f74-9566-b6ecb5a98b96\") " Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.837985 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-utilities" (OuterVolumeSpecName: "utilities") pod "dc5f2752-1355-4f74-9566-b6ecb5a98b96" (UID: "dc5f2752-1355-4f74-9566-b6ecb5a98b96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.843841 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc5f2752-1355-4f74-9566-b6ecb5a98b96-kube-api-access-dsb8v" (OuterVolumeSpecName: "kube-api-access-dsb8v") pod "dc5f2752-1355-4f74-9566-b6ecb5a98b96" (UID: "dc5f2752-1355-4f74-9566-b6ecb5a98b96"). InnerVolumeSpecName "kube-api-access-dsb8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.897517 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc5f2752-1355-4f74-9566-b6ecb5a98b96" (UID: "dc5f2752-1355-4f74-9566-b6ecb5a98b96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.939590 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.939635 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc5f2752-1355-4f74-9566-b6ecb5a98b96-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:54:26 crc kubenswrapper[4966]: I1217 09:54:26.939647 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsb8v\" (UniqueName: \"kubernetes.io/projected/dc5f2752-1355-4f74-9566-b6ecb5a98b96-kube-api-access-dsb8v\") on node \"crc\" DevicePath \"\"" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.126725 4966 generic.go:334] "Generic (PLEG): container finished" podID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerID="29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc" exitCode=0 Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.126794 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hzjvh" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.126822 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzjvh" event={"ID":"dc5f2752-1355-4f74-9566-b6ecb5a98b96","Type":"ContainerDied","Data":"29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc"} Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.128098 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hzjvh" event={"ID":"dc5f2752-1355-4f74-9566-b6ecb5a98b96","Type":"ContainerDied","Data":"ed8e21da9385fbcd6f1c9a72b7d6aac83af0dffc8a01705c8b2babcc19e0867a"} Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.128120 4966 scope.go:117] "RemoveContainer" containerID="29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.149296 4966 scope.go:117] "RemoveContainer" containerID="a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.196025 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hzjvh"] Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.209628 4966 scope.go:117] "RemoveContainer" containerID="237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.211355 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hzjvh"] Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.240465 4966 scope.go:117] "RemoveContainer" containerID="29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc" Dec 17 09:54:27 crc kubenswrapper[4966]: E1217 09:54:27.241299 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc\": container with ID starting with 29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc not found: ID does not exist" containerID="29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.241337 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc"} err="failed to get container status \"29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc\": rpc error: code = NotFound desc = could not find container \"29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc\": container with ID starting with 29d7c2833e0b0d6468aee52394b4e83519cd0a61a35775cc66b64be1d7ae3ddc not found: ID does not exist" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.241363 4966 scope.go:117] "RemoveContainer" containerID="a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467" Dec 17 09:54:27 crc kubenswrapper[4966]: E1217 09:54:27.241742 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467\": container with ID starting with a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467 not found: ID does not exist" containerID="a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.241791 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467"} err="failed to get container status \"a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467\": rpc error: code = NotFound desc = could not find container \"a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467\": container with ID starting with a2eb374c0ef07278d9cae59d5b8ade4513a6baaad40cf57c8f2f13c77fe9d467 not found: ID does not exist" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.241822 4966 scope.go:117] "RemoveContainer" containerID="237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b" Dec 17 09:54:27 crc kubenswrapper[4966]: E1217 09:54:27.242118 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b\": container with ID starting with 237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b not found: ID does not exist" containerID="237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b" Dec 17 09:54:27 crc kubenswrapper[4966]: I1217 09:54:27.242146 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b"} err="failed to get container status \"237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b\": rpc error: code = NotFound desc = could not find container \"237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b\": container with ID starting with 237f51108d489966705d7f1b3acd0d65201b86c39e112c87821dc48e567b605b not found: ID does not exist" Dec 17 09:54:28 crc kubenswrapper[4966]: I1217 09:54:28.843323 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" path="/var/lib/kubelet/pods/dc5f2752-1355-4f74-9566-b6ecb5a98b96/volumes" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.471587 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fplnh"] Dec 17 09:55:40 crc kubenswrapper[4966]: E1217 09:55:40.472869 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerName="registry-server" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.472915 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerName="registry-server" Dec 17 09:55:40 crc kubenswrapper[4966]: E1217 09:55:40.472945 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerName="extract-utilities" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.472959 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerName="extract-utilities" Dec 17 09:55:40 crc kubenswrapper[4966]: E1217 09:55:40.472988 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerName="extract-content" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.473001 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerName="extract-content" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.473300 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5f2752-1355-4f74-9566-b6ecb5a98b96" containerName="registry-server" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.475713 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.542380 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fplnh"] Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.545241 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-utilities\") pod \"community-operators-fplnh\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.545545 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-catalog-content\") pod \"community-operators-fplnh\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.545597 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmvgr\" (UniqueName: \"kubernetes.io/projected/032b6250-856b-4ded-a511-57c607a0830a-kube-api-access-qmvgr\") pod \"community-operators-fplnh\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.648996 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-catalog-content\") pod \"community-operators-fplnh\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.649420 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-catalog-content\") pod \"community-operators-fplnh\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.649792 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmvgr\" (UniqueName: \"kubernetes.io/projected/032b6250-856b-4ded-a511-57c607a0830a-kube-api-access-qmvgr\") pod \"community-operators-fplnh\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.649898 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-utilities\") pod \"community-operators-fplnh\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.650156 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-utilities\") pod \"community-operators-fplnh\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.669820 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmvgr\" (UniqueName: \"kubernetes.io/projected/032b6250-856b-4ded-a511-57c607a0830a-kube-api-access-qmvgr\") pod \"community-operators-fplnh\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:40 crc kubenswrapper[4966]: I1217 09:55:40.848150 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:41 crc kubenswrapper[4966]: I1217 09:55:41.347829 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fplnh"] Dec 17 09:55:41 crc kubenswrapper[4966]: I1217 09:55:41.806052 4966 generic.go:334] "Generic (PLEG): container finished" podID="032b6250-856b-4ded-a511-57c607a0830a" containerID="c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042" exitCode=0 Dec 17 09:55:41 crc kubenswrapper[4966]: I1217 09:55:41.806280 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fplnh" event={"ID":"032b6250-856b-4ded-a511-57c607a0830a","Type":"ContainerDied","Data":"c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042"} Dec 17 09:55:41 crc kubenswrapper[4966]: I1217 09:55:41.806340 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fplnh" event={"ID":"032b6250-856b-4ded-a511-57c607a0830a","Type":"ContainerStarted","Data":"1a237c6e7dba0ae4a27e59c93b1ef5b710d6582032ef5dbb96a5e6b58b19b215"} Dec 17 09:55:41 crc kubenswrapper[4966]: I1217 09:55:41.808376 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 09:55:43 crc kubenswrapper[4966]: I1217 09:55:43.827540 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fplnh" event={"ID":"032b6250-856b-4ded-a511-57c607a0830a","Type":"ContainerStarted","Data":"9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87"} Dec 17 09:55:44 crc kubenswrapper[4966]: I1217 09:55:44.838047 4966 generic.go:334] "Generic (PLEG): container finished" podID="032b6250-856b-4ded-a511-57c607a0830a" containerID="9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87" exitCode=0 Dec 17 09:55:44 crc kubenswrapper[4966]: I1217 09:55:44.841343 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fplnh" event={"ID":"032b6250-856b-4ded-a511-57c607a0830a","Type":"ContainerDied","Data":"9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87"} Dec 17 09:55:45 crc kubenswrapper[4966]: I1217 09:55:45.850858 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fplnh" event={"ID":"032b6250-856b-4ded-a511-57c607a0830a","Type":"ContainerStarted","Data":"38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659"} Dec 17 09:55:45 crc kubenswrapper[4966]: I1217 09:55:45.882138 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fplnh" podStartSLOduration=2.400042245 podStartE2EDuration="5.882116148s" podCreationTimestamp="2025-12-17 09:55:40 +0000 UTC" firstStartedPulling="2025-12-17 09:55:41.808018315 +0000 UTC m=+5677.353088257" lastFinishedPulling="2025-12-17 09:55:45.290092218 +0000 UTC m=+5680.835162160" observedRunningTime="2025-12-17 09:55:45.87160521 +0000 UTC m=+5681.416675172" watchObservedRunningTime="2025-12-17 09:55:45.882116148 +0000 UTC m=+5681.427186100" Dec 17 09:55:50 crc kubenswrapper[4966]: I1217 09:55:50.875159 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:50 crc kubenswrapper[4966]: I1217 09:55:50.875790 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:55:51 crc kubenswrapper[4966]: I1217 09:55:51.913447 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-fplnh" podUID="032b6250-856b-4ded-a511-57c607a0830a" containerName="registry-server" probeResult="failure" output=< Dec 17 09:55:51 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 09:55:51 crc kubenswrapper[4966]: > Dec 17 09:56:00 crc kubenswrapper[4966]: I1217 09:56:00.903791 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:56:00 crc kubenswrapper[4966]: I1217 09:56:00.954031 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:56:01 crc kubenswrapper[4966]: I1217 09:56:01.139350 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fplnh"] Dec 17 09:56:01 crc kubenswrapper[4966]: I1217 09:56:01.995614 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fplnh" podUID="032b6250-856b-4ded-a511-57c607a0830a" containerName="registry-server" containerID="cri-o://38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659" gracePeriod=2 Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.577511 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.672729 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmvgr\" (UniqueName: \"kubernetes.io/projected/032b6250-856b-4ded-a511-57c607a0830a-kube-api-access-qmvgr\") pod \"032b6250-856b-4ded-a511-57c607a0830a\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.672796 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-catalog-content\") pod \"032b6250-856b-4ded-a511-57c607a0830a\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.673315 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-utilities\") pod \"032b6250-856b-4ded-a511-57c607a0830a\" (UID: \"032b6250-856b-4ded-a511-57c607a0830a\") " Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.674419 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-utilities" (OuterVolumeSpecName: "utilities") pod "032b6250-856b-4ded-a511-57c607a0830a" (UID: "032b6250-856b-4ded-a511-57c607a0830a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.682003 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/032b6250-856b-4ded-a511-57c607a0830a-kube-api-access-qmvgr" (OuterVolumeSpecName: "kube-api-access-qmvgr") pod "032b6250-856b-4ded-a511-57c607a0830a" (UID: "032b6250-856b-4ded-a511-57c607a0830a"). InnerVolumeSpecName "kube-api-access-qmvgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.736953 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "032b6250-856b-4ded-a511-57c607a0830a" (UID: "032b6250-856b-4ded-a511-57c607a0830a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.775980 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.776019 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmvgr\" (UniqueName: \"kubernetes.io/projected/032b6250-856b-4ded-a511-57c607a0830a-kube-api-access-qmvgr\") on node \"crc\" DevicePath \"\"" Dec 17 09:56:02 crc kubenswrapper[4966]: I1217 09:56:02.776030 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032b6250-856b-4ded-a511-57c607a0830a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.005983 4966 generic.go:334] "Generic (PLEG): container finished" podID="032b6250-856b-4ded-a511-57c607a0830a" containerID="38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659" exitCode=0 Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.006030 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fplnh" event={"ID":"032b6250-856b-4ded-a511-57c607a0830a","Type":"ContainerDied","Data":"38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659"} Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.006055 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fplnh" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.006073 4966 scope.go:117] "RemoveContainer" containerID="38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.006060 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fplnh" event={"ID":"032b6250-856b-4ded-a511-57c607a0830a","Type":"ContainerDied","Data":"1a237c6e7dba0ae4a27e59c93b1ef5b710d6582032ef5dbb96a5e6b58b19b215"} Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.035298 4966 scope.go:117] "RemoveContainer" containerID="9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.035298 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fplnh"] Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.044415 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fplnh"] Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.071670 4966 scope.go:117] "RemoveContainer" containerID="c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.107509 4966 scope.go:117] "RemoveContainer" containerID="38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659" Dec 17 09:56:03 crc kubenswrapper[4966]: E1217 09:56:03.108173 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659\": container with ID starting with 38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659 not found: ID does not exist" containerID="38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.108284 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659"} err="failed to get container status \"38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659\": rpc error: code = NotFound desc = could not find container \"38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659\": container with ID starting with 38df7bff9b6d28d81f01f8e7ba143c317c25d2ba3dd2d22d84210bc56170c659 not found: ID does not exist" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.108533 4966 scope.go:117] "RemoveContainer" containerID="9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87" Dec 17 09:56:03 crc kubenswrapper[4966]: E1217 09:56:03.109149 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87\": container with ID starting with 9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87 not found: ID does not exist" containerID="9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.109196 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87"} err="failed to get container status \"9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87\": rpc error: code = NotFound desc = could not find container \"9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87\": container with ID starting with 9d4fb41dd4765c838f9e69caea5089abe85a3343f02d0c5f4c7e02e3a78fcf87 not found: ID does not exist" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.109220 4966 scope.go:117] "RemoveContainer" containerID="c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042" Dec 17 09:56:03 crc kubenswrapper[4966]: E1217 09:56:03.109843 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042\": container with ID starting with c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042 not found: ID does not exist" containerID="c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042" Dec 17 09:56:03 crc kubenswrapper[4966]: I1217 09:56:03.109887 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042"} err="failed to get container status \"c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042\": rpc error: code = NotFound desc = could not find container \"c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042\": container with ID starting with c9d72761a30e6311e84916e4981cf79bfd751bd9d14931a5ac14ecccea89b042 not found: ID does not exist" Dec 17 09:56:04 crc kubenswrapper[4966]: I1217 09:56:04.847641 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="032b6250-856b-4ded-a511-57c607a0830a" path="/var/lib/kubelet/pods/032b6250-856b-4ded-a511-57c607a0830a/volumes" Dec 17 09:56:46 crc kubenswrapper[4966]: I1217 09:56:46.808484 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:56:46 crc kubenswrapper[4966]: I1217 09:56:46.810411 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:57:16 crc kubenswrapper[4966]: I1217 09:57:16.807519 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:57:16 crc kubenswrapper[4966]: I1217 09:57:16.808072 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:57:46 crc kubenswrapper[4966]: I1217 09:57:46.807963 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 09:57:46 crc kubenswrapper[4966]: I1217 09:57:46.808388 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 09:57:46 crc kubenswrapper[4966]: I1217 09:57:46.808455 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 09:57:46 crc kubenswrapper[4966]: I1217 09:57:46.809640 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eccaf98bd2744c8937c39460a50fe0df9f527a236e1a3c4c46bd605805659ca1"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 09:57:46 crc kubenswrapper[4966]: I1217 09:57:46.809696 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://eccaf98bd2744c8937c39460a50fe0df9f527a236e1a3c4c46bd605805659ca1" gracePeriod=600 Dec 17 09:57:46 crc kubenswrapper[4966]: I1217 09:57:46.948320 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="eccaf98bd2744c8937c39460a50fe0df9f527a236e1a3c4c46bd605805659ca1" exitCode=0 Dec 17 09:57:46 crc kubenswrapper[4966]: I1217 09:57:46.948640 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"eccaf98bd2744c8937c39460a50fe0df9f527a236e1a3c4c46bd605805659ca1"} Dec 17 09:57:46 crc kubenswrapper[4966]: I1217 09:57:46.948714 4966 scope.go:117] "RemoveContainer" containerID="e7fad9507bdbd52b10ee005294048f44d8378aaa9d0914609e9ebe5144ed6dc8" Dec 17 09:57:47 crc kubenswrapper[4966]: I1217 09:57:47.963251 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad"} Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.221603 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk"] Dec 17 10:00:00 crc kubenswrapper[4966]: E1217 10:00:00.222862 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032b6250-856b-4ded-a511-57c607a0830a" containerName="extract-content" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.222905 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="032b6250-856b-4ded-a511-57c607a0830a" containerName="extract-content" Dec 17 10:00:00 crc kubenswrapper[4966]: E1217 10:00:00.222921 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032b6250-856b-4ded-a511-57c607a0830a" containerName="registry-server" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.222926 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="032b6250-856b-4ded-a511-57c607a0830a" containerName="registry-server" Dec 17 10:00:00 crc kubenswrapper[4966]: E1217 10:00:00.222934 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032b6250-856b-4ded-a511-57c607a0830a" containerName="extract-utilities" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.222941 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="032b6250-856b-4ded-a511-57c607a0830a" containerName="extract-utilities" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.223126 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="032b6250-856b-4ded-a511-57c607a0830a" containerName="registry-server" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.224565 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.231230 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.233698 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk"] Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.235016 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.392352 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-config-volume\") pod \"collect-profiles-29432760-6pzqk\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.392403 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r85zh\" (UniqueName: \"kubernetes.io/projected/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-kube-api-access-r85zh\") pod \"collect-profiles-29432760-6pzqk\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.392476 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-secret-volume\") pod \"collect-profiles-29432760-6pzqk\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.494350 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-config-volume\") pod \"collect-profiles-29432760-6pzqk\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.494401 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r85zh\" (UniqueName: \"kubernetes.io/projected/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-kube-api-access-r85zh\") pod \"collect-profiles-29432760-6pzqk\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.494490 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-secret-volume\") pod \"collect-profiles-29432760-6pzqk\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.496156 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-config-volume\") pod \"collect-profiles-29432760-6pzqk\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.510270 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-secret-volume\") pod \"collect-profiles-29432760-6pzqk\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.517798 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r85zh\" (UniqueName: \"kubernetes.io/projected/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-kube-api-access-r85zh\") pod \"collect-profiles-29432760-6pzqk\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:00 crc kubenswrapper[4966]: I1217 10:00:00.558651 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:01 crc kubenswrapper[4966]: I1217 10:00:01.066090 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk"] Dec 17 10:00:01 crc kubenswrapper[4966]: I1217 10:00:01.182133 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" event={"ID":"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b","Type":"ContainerStarted","Data":"8788014ab2acbfb5517dcca57c126d1ff0a67a185ffac5f242f4c98cb36a3e66"} Dec 17 10:00:01 crc kubenswrapper[4966]: E1217 10:00:01.900546 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda27a418b_f81c_4546_9e7f_f0c4a8d1ae0b.slice/crio-conmon-6756d56218b2294c5a2fe45c00185ff724671f2d1305b390694d00893dd34b26.scope\": RecentStats: unable to find data in memory cache]" Dec 17 10:00:02 crc kubenswrapper[4966]: I1217 10:00:02.202106 4966 generic.go:334] "Generic (PLEG): container finished" podID="a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b" containerID="6756d56218b2294c5a2fe45c00185ff724671f2d1305b390694d00893dd34b26" exitCode=0 Dec 17 10:00:02 crc kubenswrapper[4966]: I1217 10:00:02.202466 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" event={"ID":"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b","Type":"ContainerDied","Data":"6756d56218b2294c5a2fe45c00185ff724671f2d1305b390694d00893dd34b26"} Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.598912 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.765043 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r85zh\" (UniqueName: \"kubernetes.io/projected/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-kube-api-access-r85zh\") pod \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.765083 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-config-volume\") pod \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.765146 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-secret-volume\") pod \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\" (UID: \"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b\") " Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.766414 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-config-volume" (OuterVolumeSpecName: "config-volume") pod "a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b" (UID: "a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.770990 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b" (UID: "a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.771214 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-kube-api-access-r85zh" (OuterVolumeSpecName: "kube-api-access-r85zh") pod "a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b" (UID: "a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b"). InnerVolumeSpecName "kube-api-access-r85zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.867734 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r85zh\" (UniqueName: \"kubernetes.io/projected/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-kube-api-access-r85zh\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.867776 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:03 crc kubenswrapper[4966]: I1217 10:00:03.867788 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:04 crc kubenswrapper[4966]: I1217 10:00:04.224937 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" event={"ID":"a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b","Type":"ContainerDied","Data":"8788014ab2acbfb5517dcca57c126d1ff0a67a185ffac5f242f4c98cb36a3e66"} Dec 17 10:00:04 crc kubenswrapper[4966]: I1217 10:00:04.225984 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8788014ab2acbfb5517dcca57c126d1ff0a67a185ffac5f242f4c98cb36a3e66" Dec 17 10:00:04 crc kubenswrapper[4966]: I1217 10:00:04.225058 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk" Dec 17 10:00:04 crc kubenswrapper[4966]: I1217 10:00:04.689018 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd"] Dec 17 10:00:04 crc kubenswrapper[4966]: I1217 10:00:04.698049 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432715-p66xd"] Dec 17 10:00:04 crc kubenswrapper[4966]: I1217 10:00:04.848907 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75b6bdc9-e860-4378-97cc-14c90d2b6e6f" path="/var/lib/kubelet/pods/75b6bdc9-e860-4378-97cc-14c90d2b6e6f/volumes" Dec 17 10:00:05 crc kubenswrapper[4966]: I1217 10:00:05.234902 4966 generic.go:334] "Generic (PLEG): container finished" podID="780b74fd-a690-416e-80f1-3e28a75bd67c" containerID="0d516a9cf423ca68ecb9ba8e34eb110267cd67f48464edd92b0b1ce1a9e0a2f0" exitCode=0 Dec 17 10:00:05 crc kubenswrapper[4966]: I1217 10:00:05.234943 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"780b74fd-a690-416e-80f1-3e28a75bd67c","Type":"ContainerDied","Data":"0d516a9cf423ca68ecb9ba8e34eb110267cd67f48464edd92b0b1ce1a9e0a2f0"} Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.809680 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.835187 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-workdir\") pod \"780b74fd-a690-416e-80f1-3e28a75bd67c\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.835280 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-temporary\") pod \"780b74fd-a690-416e-80f1-3e28a75bd67c\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.835338 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config\") pod \"780b74fd-a690-416e-80f1-3e28a75bd67c\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.835370 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ssh-key\") pod \"780b74fd-a690-416e-80f1-3e28a75bd67c\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.835528 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"780b74fd-a690-416e-80f1-3e28a75bd67c\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.835604 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ca-certs\") pod \"780b74fd-a690-416e-80f1-3e28a75bd67c\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.835674 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2d4r\" (UniqueName: \"kubernetes.io/projected/780b74fd-a690-416e-80f1-3e28a75bd67c-kube-api-access-g2d4r\") pod \"780b74fd-a690-416e-80f1-3e28a75bd67c\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.835701 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config-secret\") pod \"780b74fd-a690-416e-80f1-3e28a75bd67c\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.835736 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-config-data\") pod \"780b74fd-a690-416e-80f1-3e28a75bd67c\" (UID: \"780b74fd-a690-416e-80f1-3e28a75bd67c\") " Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.838633 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "780b74fd-a690-416e-80f1-3e28a75bd67c" (UID: "780b74fd-a690-416e-80f1-3e28a75bd67c"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.839130 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-config-data" (OuterVolumeSpecName: "config-data") pod "780b74fd-a690-416e-80f1-3e28a75bd67c" (UID: "780b74fd-a690-416e-80f1-3e28a75bd67c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.880103 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/780b74fd-a690-416e-80f1-3e28a75bd67c-kube-api-access-g2d4r" (OuterVolumeSpecName: "kube-api-access-g2d4r") pod "780b74fd-a690-416e-80f1-3e28a75bd67c" (UID: "780b74fd-a690-416e-80f1-3e28a75bd67c"). InnerVolumeSpecName "kube-api-access-g2d4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.918676 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "780b74fd-a690-416e-80f1-3e28a75bd67c" (UID: "780b74fd-a690-416e-80f1-3e28a75bd67c"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.941693 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "780b74fd-a690-416e-80f1-3e28a75bd67c" (UID: "780b74fd-a690-416e-80f1-3e28a75bd67c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.943711 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.943732 4966 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.943743 4966 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/780b74fd-a690-416e-80f1-3e28a75bd67c-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.943752 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.943761 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2d4r\" (UniqueName: \"kubernetes.io/projected/780b74fd-a690-416e-80f1-3e28a75bd67c-kube-api-access-g2d4r\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.950001 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "780b74fd-a690-416e-80f1-3e28a75bd67c" (UID: "780b74fd-a690-416e-80f1-3e28a75bd67c"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.964190 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-thread-testing"] Dec 17 10:00:06 crc kubenswrapper[4966]: E1217 10:00:06.964638 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="780b74fd-a690-416e-80f1-3e28a75bd67c" containerName="tempest-tests-tempest-tests-runner" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.964656 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="780b74fd-a690-416e-80f1-3e28a75bd67c" containerName="tempest-tests-tempest-tests-runner" Dec 17 10:00:06 crc kubenswrapper[4966]: E1217 10:00:06.964699 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b" containerName="collect-profiles" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.964706 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b" containerName="collect-profiles" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.964924 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b" containerName="collect-profiles" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.964940 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="780b74fd-a690-416e-80f1-3e28a75bd67c" containerName="tempest-tests-tempest-tests-runner" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.965556 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.967944 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.968019 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.968615 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "780b74fd-a690-416e-80f1-3e28a75bd67c" (UID: "780b74fd-a690-416e-80f1-3e28a75bd67c"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.985015 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "780b74fd-a690-416e-80f1-3e28a75bd67c" (UID: "780b74fd-a690-416e-80f1-3e28a75bd67c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:00:06 crc kubenswrapper[4966]: I1217 10:00:06.995219 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-thread-testing"] Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.015368 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "780b74fd-a690-416e-80f1-3e28a75bd67c" (UID: "780b74fd-a690-416e-80f1-3e28a75bd67c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.044551 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ssh-key\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.044866 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.044986 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.045066 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-config-data\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.045170 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ca-certs\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.045319 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.045404 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.045428 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.045539 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8445b\" (UniqueName: \"kubernetes.io/projected/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-kube-api-access-8445b\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.045684 4966 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.045701 4966 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.045712 4966 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/780b74fd-a690-416e-80f1-3e28a75bd67c-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.071758 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.147507 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.147567 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.147588 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.147678 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8445b\" (UniqueName: \"kubernetes.io/projected/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-kube-api-access-8445b\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.147739 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ssh-key\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.148344 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.148562 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.148639 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-config-data\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.149013 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ca-certs\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.150007 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-config-data\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.150026 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.150672 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.152052 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ssh-key\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.155711 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.156139 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ca-certs\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.172667 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8445b\" (UniqueName: \"kubernetes.io/projected/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-kube-api-access-8445b\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.261052 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"780b74fd-a690-416e-80f1-3e28a75bd67c","Type":"ContainerDied","Data":"a2cc6431d20ac839b5a4426d613ec41d80f5428efd6057cb2523ad49ccf32aae"} Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.261383 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2cc6431d20ac839b5a4426d613ec41d80f5428efd6057cb2523ad49ccf32aae" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.261293 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.366908 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 10:00:07 crc kubenswrapper[4966]: I1217 10:00:07.990134 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-thread-testing"] Dec 17 10:00:08 crc kubenswrapper[4966]: I1217 10:00:08.269844 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"efdaa7df-eeb5-48ec-afb5-a6c5cff26269","Type":"ContainerStarted","Data":"75f1df2d6e44e8ecea0dc9c0c2f7460793b15266299e63f02accb7db09e034fd"} Dec 17 10:00:10 crc kubenswrapper[4966]: I1217 10:00:10.297135 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"efdaa7df-eeb5-48ec-afb5-a6c5cff26269","Type":"ContainerStarted","Data":"315a71f8fedc951dad75a270b142fbee024b98cfa6c046dcc1ced49081f4486d"} Dec 17 10:00:16 crc kubenswrapper[4966]: I1217 10:00:16.808224 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:00:16 crc kubenswrapper[4966]: I1217 10:00:16.808822 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:00:46 crc kubenswrapper[4966]: I1217 10:00:46.807410 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:00:46 crc kubenswrapper[4966]: I1217 10:00:46.808016 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:00:59 crc kubenswrapper[4966]: I1217 10:00:59.967772 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" podStartSLOduration=53.967752764 podStartE2EDuration="53.967752764s" podCreationTimestamp="2025-12-17 10:00:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 10:00:10.324355969 +0000 UTC m=+5945.869425911" watchObservedRunningTime="2025-12-17 10:00:59.967752764 +0000 UTC m=+5995.512822716" Dec 17 10:00:59 crc kubenswrapper[4966]: I1217 10:00:59.970689 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-545d74fd97-8brzm"] Dec 17 10:00:59 crc kubenswrapper[4966]: I1217 10:00:59.972927 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:00:59 crc kubenswrapper[4966]: I1217 10:00:59.989649 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-545d74fd97-8brzm"] Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.149275 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29432761-w4pqp"] Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.150685 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.164637 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-internal-tls-certs\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.164736 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-public-tls-certs\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.164779 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-httpd-config\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.164807 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-ovndb-tls-certs\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.164842 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-combined-ca-bundle\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.164899 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr6rm\" (UniqueName: \"kubernetes.io/projected/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-kube-api-access-qr6rm\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.164934 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-config\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.165149 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29432761-w4pqp"] Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.182735 4966 scope.go:117] "RemoveContainer" containerID="dad663132e96f3e6f49d634e677d2bda89a09662e133194bf9b51c20c54be440" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.266543 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-public-tls-certs\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.267182 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-httpd-config\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.267326 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-ovndb-tls-certs\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.267490 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-combined-ca-bundle\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.267692 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb6mh\" (UniqueName: \"kubernetes.io/projected/25c43855-7134-43e5-9ff0-e37bc27a03c4-kube-api-access-fb6mh\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.267909 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr6rm\" (UniqueName: \"kubernetes.io/projected/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-kube-api-access-qr6rm\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.268033 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-combined-ca-bundle\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.268187 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-config\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.268327 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-fernet-keys\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.268575 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-internal-tls-certs\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.269028 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-config-data\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.277927 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-ovndb-tls-certs\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.278768 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-public-tls-certs\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.280325 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-config\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.286405 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-combined-ca-bundle\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.286519 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-httpd-config\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.290074 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-internal-tls-certs\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.292532 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr6rm\" (UniqueName: \"kubernetes.io/projected/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-kube-api-access-qr6rm\") pod \"neutron-545d74fd97-8brzm\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.294451 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.371335 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-combined-ca-bundle\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.371847 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-fernet-keys\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.372038 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-config-data\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.372206 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb6mh\" (UniqueName: \"kubernetes.io/projected/25c43855-7134-43e5-9ff0-e37bc27a03c4-kube-api-access-fb6mh\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.376327 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-fernet-keys\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.377048 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-combined-ca-bundle\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.397735 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb6mh\" (UniqueName: \"kubernetes.io/projected/25c43855-7134-43e5-9ff0-e37bc27a03c4-kube-api-access-fb6mh\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.409664 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-config-data\") pod \"keystone-cron-29432761-w4pqp\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.470744 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:00 crc kubenswrapper[4966]: I1217 10:01:00.919517 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-545d74fd97-8brzm"] Dec 17 10:01:01 crc kubenswrapper[4966]: I1217 10:01:01.014110 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29432761-w4pqp"] Dec 17 10:01:01 crc kubenswrapper[4966]: W1217 10:01:01.017553 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25c43855_7134_43e5_9ff0_e37bc27a03c4.slice/crio-1e3c03051f25b2dcfd6e7a76a1ec1b090c7ec95897f7bba772f1d34cbb1b75b6 WatchSource:0}: Error finding container 1e3c03051f25b2dcfd6e7a76a1ec1b090c7ec95897f7bba772f1d34cbb1b75b6: Status 404 returned error can't find the container with id 1e3c03051f25b2dcfd6e7a76a1ec1b090c7ec95897f7bba772f1d34cbb1b75b6 Dec 17 10:01:01 crc kubenswrapper[4966]: I1217 10:01:01.800077 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432761-w4pqp" event={"ID":"25c43855-7134-43e5-9ff0-e37bc27a03c4","Type":"ContainerStarted","Data":"8ba0c78900d5db45951a2ca72c73fc007030d350f07dfdb9019b81fc175d2b37"} Dec 17 10:01:01 crc kubenswrapper[4966]: I1217 10:01:01.800379 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432761-w4pqp" event={"ID":"25c43855-7134-43e5-9ff0-e37bc27a03c4","Type":"ContainerStarted","Data":"1e3c03051f25b2dcfd6e7a76a1ec1b090c7ec95897f7bba772f1d34cbb1b75b6"} Dec 17 10:01:01 crc kubenswrapper[4966]: I1217 10:01:01.806015 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-545d74fd97-8brzm" event={"ID":"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9","Type":"ContainerStarted","Data":"770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08"} Dec 17 10:01:01 crc kubenswrapper[4966]: I1217 10:01:01.806057 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-545d74fd97-8brzm" event={"ID":"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9","Type":"ContainerStarted","Data":"bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb"} Dec 17 10:01:01 crc kubenswrapper[4966]: I1217 10:01:01.806067 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-545d74fd97-8brzm" event={"ID":"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9","Type":"ContainerStarted","Data":"4e267a69e5de66f2ee6287defc4bd12204333330f6627e0817988675e63fb79e"} Dec 17 10:01:01 crc kubenswrapper[4966]: I1217 10:01:01.806178 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:01 crc kubenswrapper[4966]: I1217 10:01:01.823087 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29432761-w4pqp" podStartSLOduration=1.823068604 podStartE2EDuration="1.823068604s" podCreationTimestamp="2025-12-17 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 10:01:01.816481135 +0000 UTC m=+5997.361551077" watchObservedRunningTime="2025-12-17 10:01:01.823068604 +0000 UTC m=+5997.368138546" Dec 17 10:01:01 crc kubenswrapper[4966]: I1217 10:01:01.867927 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-545d74fd97-8brzm" podStartSLOduration=2.867860116 podStartE2EDuration="2.867860116s" podCreationTimestamp="2025-12-17 10:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 10:01:01.842457003 +0000 UTC m=+5997.387526965" watchObservedRunningTime="2025-12-17 10:01:01.867860116 +0000 UTC m=+5997.412930058" Dec 17 10:01:03 crc kubenswrapper[4966]: I1217 10:01:03.836095 4966 generic.go:334] "Generic (PLEG): container finished" podID="25c43855-7134-43e5-9ff0-e37bc27a03c4" containerID="8ba0c78900d5db45951a2ca72c73fc007030d350f07dfdb9019b81fc175d2b37" exitCode=0 Dec 17 10:01:03 crc kubenswrapper[4966]: I1217 10:01:03.836138 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432761-w4pqp" event={"ID":"25c43855-7134-43e5-9ff0-e37bc27a03c4","Type":"ContainerDied","Data":"8ba0c78900d5db45951a2ca72c73fc007030d350f07dfdb9019b81fc175d2b37"} Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.228229 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.290505 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb6mh\" (UniqueName: \"kubernetes.io/projected/25c43855-7134-43e5-9ff0-e37bc27a03c4-kube-api-access-fb6mh\") pod \"25c43855-7134-43e5-9ff0-e37bc27a03c4\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.290721 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-combined-ca-bundle\") pod \"25c43855-7134-43e5-9ff0-e37bc27a03c4\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.290827 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-config-data\") pod \"25c43855-7134-43e5-9ff0-e37bc27a03c4\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.290888 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-fernet-keys\") pod \"25c43855-7134-43e5-9ff0-e37bc27a03c4\" (UID: \"25c43855-7134-43e5-9ff0-e37bc27a03c4\") " Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.302306 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "25c43855-7134-43e5-9ff0-e37bc27a03c4" (UID: "25c43855-7134-43e5-9ff0-e37bc27a03c4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.312154 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25c43855-7134-43e5-9ff0-e37bc27a03c4-kube-api-access-fb6mh" (OuterVolumeSpecName: "kube-api-access-fb6mh") pod "25c43855-7134-43e5-9ff0-e37bc27a03c4" (UID: "25c43855-7134-43e5-9ff0-e37bc27a03c4"). InnerVolumeSpecName "kube-api-access-fb6mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.327993 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25c43855-7134-43e5-9ff0-e37bc27a03c4" (UID: "25c43855-7134-43e5-9ff0-e37bc27a03c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.351117 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-config-data" (OuterVolumeSpecName: "config-data") pod "25c43855-7134-43e5-9ff0-e37bc27a03c4" (UID: "25c43855-7134-43e5-9ff0-e37bc27a03c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.393460 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb6mh\" (UniqueName: \"kubernetes.io/projected/25c43855-7134-43e5-9ff0-e37bc27a03c4-kube-api-access-fb6mh\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.393494 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.393504 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.393515 4966 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25c43855-7134-43e5-9ff0-e37bc27a03c4-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.862004 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432761-w4pqp" event={"ID":"25c43855-7134-43e5-9ff0-e37bc27a03c4","Type":"ContainerDied","Data":"1e3c03051f25b2dcfd6e7a76a1ec1b090c7ec95897f7bba772f1d34cbb1b75b6"} Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.862289 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e3c03051f25b2dcfd6e7a76a1ec1b090c7ec95897f7bba772f1d34cbb1b75b6" Dec 17 10:01:05 crc kubenswrapper[4966]: I1217 10:01:05.862081 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432761-w4pqp" Dec 17 10:01:16 crc kubenswrapper[4966]: I1217 10:01:16.808391 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:01:16 crc kubenswrapper[4966]: I1217 10:01:16.808960 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:01:16 crc kubenswrapper[4966]: I1217 10:01:16.809021 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:01:16 crc kubenswrapper[4966]: I1217 10:01:16.810318 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:01:16 crc kubenswrapper[4966]: I1217 10:01:16.810380 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" gracePeriod=600 Dec 17 10:01:16 crc kubenswrapper[4966]: E1217 10:01:16.941776 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:01:17 crc kubenswrapper[4966]: I1217 10:01:17.024825 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" exitCode=0 Dec 17 10:01:17 crc kubenswrapper[4966]: I1217 10:01:17.024868 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad"} Dec 17 10:01:17 crc kubenswrapper[4966]: I1217 10:01:17.024933 4966 scope.go:117] "RemoveContainer" containerID="eccaf98bd2744c8937c39460a50fe0df9f527a236e1a3c4c46bd605805659ca1" Dec 17 10:01:17 crc kubenswrapper[4966]: I1217 10:01:17.025586 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:01:17 crc kubenswrapper[4966]: E1217 10:01:17.025929 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:01:29 crc kubenswrapper[4966]: I1217 10:01:29.830967 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:01:29 crc kubenswrapper[4966]: E1217 10:01:29.831835 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:01:30 crc kubenswrapper[4966]: I1217 10:01:30.313803 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:01:30 crc kubenswrapper[4966]: I1217 10:01:30.406008 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f575c747c-8xn4b"] Dec 17 10:01:30 crc kubenswrapper[4966]: I1217 10:01:30.406724 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f575c747c-8xn4b" podUID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerName="neutron-api" containerID="cri-o://dcb9949ed79a7c30f72f4516d42dbe7ad4d4d56abffe4c13a496c110fc4c0627" gracePeriod=30 Dec 17 10:01:30 crc kubenswrapper[4966]: I1217 10:01:30.406804 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f575c747c-8xn4b" podUID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerName="neutron-httpd" containerID="cri-o://58cc81c70881d77dbf0be09d4674cbc1ce6bb83b337e24017bece572a468bdfa" gracePeriod=30 Dec 17 10:01:31 crc kubenswrapper[4966]: I1217 10:01:31.163644 4966 generic.go:334] "Generic (PLEG): container finished" podID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerID="58cc81c70881d77dbf0be09d4674cbc1ce6bb83b337e24017bece572a468bdfa" exitCode=0 Dec 17 10:01:31 crc kubenswrapper[4966]: I1217 10:01:31.163689 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f575c747c-8xn4b" event={"ID":"bdefbb9f-3322-4e19-98bb-ab6c11b00eec","Type":"ContainerDied","Data":"58cc81c70881d77dbf0be09d4674cbc1ce6bb83b337e24017bece572a468bdfa"} Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.233221 4966 generic.go:334] "Generic (PLEG): container finished" podID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerID="dcb9949ed79a7c30f72f4516d42dbe7ad4d4d56abffe4c13a496c110fc4c0627" exitCode=0 Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.233294 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f575c747c-8xn4b" event={"ID":"bdefbb9f-3322-4e19-98bb-ab6c11b00eec","Type":"ContainerDied","Data":"dcb9949ed79a7c30f72f4516d42dbe7ad4d4d56abffe4c13a496c110fc4c0627"} Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.617668 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.759569 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-combined-ca-bundle\") pod \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.759663 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-public-tls-certs\") pod \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.759837 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-internal-tls-certs\") pod \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.759857 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-httpd-config\") pod \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.759951 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htcd9\" (UniqueName: \"kubernetes.io/projected/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-kube-api-access-htcd9\") pod \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.759973 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-ovndb-tls-certs\") pod \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.759991 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-config\") pod \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.799747 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "bdefbb9f-3322-4e19-98bb-ab6c11b00eec" (UID: "bdefbb9f-3322-4e19-98bb-ab6c11b00eec"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.828114 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-kube-api-access-htcd9" (OuterVolumeSpecName: "kube-api-access-htcd9") pod "bdefbb9f-3322-4e19-98bb-ab6c11b00eec" (UID: "bdefbb9f-3322-4e19-98bb-ab6c11b00eec"). InnerVolumeSpecName "kube-api-access-htcd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.873168 4966 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.873209 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htcd9\" (UniqueName: \"kubernetes.io/projected/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-kube-api-access-htcd9\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.927221 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdefbb9f-3322-4e19-98bb-ab6c11b00eec" (UID: "bdefbb9f-3322-4e19-98bb-ab6c11b00eec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.942057 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bdefbb9f-3322-4e19-98bb-ab6c11b00eec" (UID: "bdefbb9f-3322-4e19-98bb-ab6c11b00eec"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.955485 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-config" (OuterVolumeSpecName: "config") pod "bdefbb9f-3322-4e19-98bb-ab6c11b00eec" (UID: "bdefbb9f-3322-4e19-98bb-ab6c11b00eec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.973723 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bdefbb9f-3322-4e19-98bb-ab6c11b00eec" (UID: "bdefbb9f-3322-4e19-98bb-ab6c11b00eec"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.974449 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-internal-tls-certs\") pod \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\" (UID: \"bdefbb9f-3322-4e19-98bb-ab6c11b00eec\") " Dec 17 10:01:38 crc kubenswrapper[4966]: W1217 10:01:38.976058 4966 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/bdefbb9f-3322-4e19-98bb-ab6c11b00eec/volumes/kubernetes.io~secret/internal-tls-certs Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.976086 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bdefbb9f-3322-4e19-98bb-ab6c11b00eec" (UID: "bdefbb9f-3322-4e19-98bb-ab6c11b00eec"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.977239 4966 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.977293 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-config\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.977357 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:38 crc kubenswrapper[4966]: I1217 10:01:38.977371 4966 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:39 crc kubenswrapper[4966]: I1217 10:01:39.000820 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "bdefbb9f-3322-4e19-98bb-ab6c11b00eec" (UID: "bdefbb9f-3322-4e19-98bb-ab6c11b00eec"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:01:39 crc kubenswrapper[4966]: I1217 10:01:39.078226 4966 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdefbb9f-3322-4e19-98bb-ab6c11b00eec-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 10:01:39 crc kubenswrapper[4966]: I1217 10:01:39.244008 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f575c747c-8xn4b" event={"ID":"bdefbb9f-3322-4e19-98bb-ab6c11b00eec","Type":"ContainerDied","Data":"2f3055015c2555bf29cc516696fd3ed3eda444d57a36699d7aca76d64dde8dcf"} Dec 17 10:01:39 crc kubenswrapper[4966]: I1217 10:01:39.244068 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f575c747c-8xn4b" Dec 17 10:01:39 crc kubenswrapper[4966]: I1217 10:01:39.244333 4966 scope.go:117] "RemoveContainer" containerID="58cc81c70881d77dbf0be09d4674cbc1ce6bb83b337e24017bece572a468bdfa" Dec 17 10:01:39 crc kubenswrapper[4966]: I1217 10:01:39.278151 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f575c747c-8xn4b"] Dec 17 10:01:39 crc kubenswrapper[4966]: I1217 10:01:39.713680 4966 scope.go:117] "RemoveContainer" containerID="dcb9949ed79a7c30f72f4516d42dbe7ad4d4d56abffe4c13a496c110fc4c0627" Dec 17 10:01:39 crc kubenswrapper[4966]: I1217 10:01:39.714526 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7f575c747c-8xn4b"] Dec 17 10:01:40 crc kubenswrapper[4966]: I1217 10:01:40.831251 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:01:40 crc kubenswrapper[4966]: E1217 10:01:40.832144 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:01:40 crc kubenswrapper[4966]: I1217 10:01:40.846407 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" path="/var/lib/kubelet/pods/bdefbb9f-3322-4e19-98bb-ab6c11b00eec/volumes" Dec 17 10:01:55 crc kubenswrapper[4966]: I1217 10:01:55.831199 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:01:55 crc kubenswrapper[4966]: E1217 10:01:55.832083 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:02:10 crc kubenswrapper[4966]: I1217 10:02:10.830918 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:02:10 crc kubenswrapper[4966]: E1217 10:02:10.831676 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:02:25 crc kubenswrapper[4966]: I1217 10:02:25.830865 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:02:25 crc kubenswrapper[4966]: E1217 10:02:25.833095 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:02:39 crc kubenswrapper[4966]: I1217 10:02:39.830536 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:02:39 crc kubenswrapper[4966]: E1217 10:02:39.831259 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.654733 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fvf5s"] Dec 17 10:02:47 crc kubenswrapper[4966]: E1217 10:02:47.655901 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerName="neutron-api" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.655917 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerName="neutron-api" Dec 17 10:02:47 crc kubenswrapper[4966]: E1217 10:02:47.655935 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c43855-7134-43e5-9ff0-e37bc27a03c4" containerName="keystone-cron" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.655943 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c43855-7134-43e5-9ff0-e37bc27a03c4" containerName="keystone-cron" Dec 17 10:02:47 crc kubenswrapper[4966]: E1217 10:02:47.655960 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerName="neutron-httpd" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.655968 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerName="neutron-httpd" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.656243 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerName="neutron-httpd" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.656259 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="25c43855-7134-43e5-9ff0-e37bc27a03c4" containerName="keystone-cron" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.656288 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdefbb9f-3322-4e19-98bb-ab6c11b00eec" containerName="neutron-api" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.660720 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.683120 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fvf5s"] Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.743993 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztjkn\" (UniqueName: \"kubernetes.io/projected/97effbb7-ea20-49ec-a328-182815c3c087-kube-api-access-ztjkn\") pod \"redhat-operators-fvf5s\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.744322 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-catalog-content\") pod \"redhat-operators-fvf5s\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.744394 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-utilities\") pod \"redhat-operators-fvf5s\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.847399 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-catalog-content\") pod \"redhat-operators-fvf5s\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.848084 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-catalog-content\") pod \"redhat-operators-fvf5s\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.848320 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-utilities\") pod \"redhat-operators-fvf5s\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.848543 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztjkn\" (UniqueName: \"kubernetes.io/projected/97effbb7-ea20-49ec-a328-182815c3c087-kube-api-access-ztjkn\") pod \"redhat-operators-fvf5s\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.848642 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-utilities\") pod \"redhat-operators-fvf5s\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.873498 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztjkn\" (UniqueName: \"kubernetes.io/projected/97effbb7-ea20-49ec-a328-182815c3c087-kube-api-access-ztjkn\") pod \"redhat-operators-fvf5s\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:47 crc kubenswrapper[4966]: I1217 10:02:47.992656 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:48 crc kubenswrapper[4966]: I1217 10:02:48.505727 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fvf5s"] Dec 17 10:02:48 crc kubenswrapper[4966]: I1217 10:02:48.860898 4966 generic.go:334] "Generic (PLEG): container finished" podID="97effbb7-ea20-49ec-a328-182815c3c087" containerID="de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8" exitCode=0 Dec 17 10:02:48 crc kubenswrapper[4966]: I1217 10:02:48.860964 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvf5s" event={"ID":"97effbb7-ea20-49ec-a328-182815c3c087","Type":"ContainerDied","Data":"de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8"} Dec 17 10:02:48 crc kubenswrapper[4966]: I1217 10:02:48.860991 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvf5s" event={"ID":"97effbb7-ea20-49ec-a328-182815c3c087","Type":"ContainerStarted","Data":"d87c0b9550357885eacaaedf7a8bdb2f615818af83adb684dc245d46a794c92a"} Dec 17 10:02:48 crc kubenswrapper[4966]: I1217 10:02:48.863802 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 10:02:49 crc kubenswrapper[4966]: I1217 10:02:49.882083 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvf5s" event={"ID":"97effbb7-ea20-49ec-a328-182815c3c087","Type":"ContainerStarted","Data":"d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646"} Dec 17 10:02:51 crc kubenswrapper[4966]: I1217 10:02:51.830945 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:02:51 crc kubenswrapper[4966]: E1217 10:02:51.831430 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:02:53 crc kubenswrapper[4966]: I1217 10:02:53.923658 4966 generic.go:334] "Generic (PLEG): container finished" podID="97effbb7-ea20-49ec-a328-182815c3c087" containerID="d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646" exitCode=0 Dec 17 10:02:53 crc kubenswrapper[4966]: I1217 10:02:53.923728 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvf5s" event={"ID":"97effbb7-ea20-49ec-a328-182815c3c087","Type":"ContainerDied","Data":"d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646"} Dec 17 10:02:54 crc kubenswrapper[4966]: I1217 10:02:54.938320 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvf5s" event={"ID":"97effbb7-ea20-49ec-a328-182815c3c087","Type":"ContainerStarted","Data":"36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff"} Dec 17 10:02:54 crc kubenswrapper[4966]: I1217 10:02:54.976779 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fvf5s" podStartSLOduration=2.379776604 podStartE2EDuration="7.976752947s" podCreationTimestamp="2025-12-17 10:02:47 +0000 UTC" firstStartedPulling="2025-12-17 10:02:48.862920695 +0000 UTC m=+6104.407990637" lastFinishedPulling="2025-12-17 10:02:54.459897038 +0000 UTC m=+6110.004966980" observedRunningTime="2025-12-17 10:02:54.969499699 +0000 UTC m=+6110.514569661" watchObservedRunningTime="2025-12-17 10:02:54.976752947 +0000 UTC m=+6110.521822899" Dec 17 10:02:57 crc kubenswrapper[4966]: I1217 10:02:57.993402 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:57 crc kubenswrapper[4966]: I1217 10:02:57.993710 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:02:59 crc kubenswrapper[4966]: I1217 10:02:59.053633 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fvf5s" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="registry-server" probeResult="failure" output=< Dec 17 10:02:59 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:02:59 crc kubenswrapper[4966]: > Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.451027 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fm2ld"] Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.454055 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.475356 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fm2ld"] Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.526545 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-utilities\") pod \"redhat-marketplace-fm2ld\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.526604 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-catalog-content\") pod \"redhat-marketplace-fm2ld\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.526661 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs995\" (UniqueName: \"kubernetes.io/projected/c141723b-23b4-49a3-b743-678314e478e5-kube-api-access-rs995\") pod \"redhat-marketplace-fm2ld\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.627489 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-catalog-content\") pod \"redhat-marketplace-fm2ld\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.627820 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs995\" (UniqueName: \"kubernetes.io/projected/c141723b-23b4-49a3-b743-678314e478e5-kube-api-access-rs995\") pod \"redhat-marketplace-fm2ld\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.627970 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-catalog-content\") pod \"redhat-marketplace-fm2ld\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.628269 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-utilities\") pod \"redhat-marketplace-fm2ld\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.628535 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-utilities\") pod \"redhat-marketplace-fm2ld\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.667682 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs995\" (UniqueName: \"kubernetes.io/projected/c141723b-23b4-49a3-b743-678314e478e5-kube-api-access-rs995\") pod \"redhat-marketplace-fm2ld\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.772287 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:02 crc kubenswrapper[4966]: I1217 10:03:02.831918 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:03:02 crc kubenswrapper[4966]: E1217 10:03:02.832511 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:03:03 crc kubenswrapper[4966]: I1217 10:03:03.370088 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fm2ld"] Dec 17 10:03:03 crc kubenswrapper[4966]: W1217 10:03:03.372096 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc141723b_23b4_49a3_b743_678314e478e5.slice/crio-a73a1bfd8efdbecafe90468e871c3655637f6425afed01643d68dcc7ebf9c1ce WatchSource:0}: Error finding container a73a1bfd8efdbecafe90468e871c3655637f6425afed01643d68dcc7ebf9c1ce: Status 404 returned error can't find the container with id a73a1bfd8efdbecafe90468e871c3655637f6425afed01643d68dcc7ebf9c1ce Dec 17 10:03:04 crc kubenswrapper[4966]: I1217 10:03:04.022327 4966 generic.go:334] "Generic (PLEG): container finished" podID="c141723b-23b4-49a3-b743-678314e478e5" containerID="7686d79bd9e6ff9e7d53a0a80a37a1cdc05d6b4f80a314b6975d201f3a0daf5a" exitCode=0 Dec 17 10:03:04 crc kubenswrapper[4966]: I1217 10:03:04.022501 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fm2ld" event={"ID":"c141723b-23b4-49a3-b743-678314e478e5","Type":"ContainerDied","Data":"7686d79bd9e6ff9e7d53a0a80a37a1cdc05d6b4f80a314b6975d201f3a0daf5a"} Dec 17 10:03:04 crc kubenswrapper[4966]: I1217 10:03:04.022617 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fm2ld" event={"ID":"c141723b-23b4-49a3-b743-678314e478e5","Type":"ContainerStarted","Data":"a73a1bfd8efdbecafe90468e871c3655637f6425afed01643d68dcc7ebf9c1ce"} Dec 17 10:03:06 crc kubenswrapper[4966]: I1217 10:03:06.045781 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fm2ld" event={"ID":"c141723b-23b4-49a3-b743-678314e478e5","Type":"ContainerDied","Data":"71d6fe931446bcb2ecec99f85415841efb1e65fcc7391b4960313e5d9df5dd7c"} Dec 17 10:03:06 crc kubenswrapper[4966]: I1217 10:03:06.046956 4966 generic.go:334] "Generic (PLEG): container finished" podID="c141723b-23b4-49a3-b743-678314e478e5" containerID="71d6fe931446bcb2ecec99f85415841efb1e65fcc7391b4960313e5d9df5dd7c" exitCode=0 Dec 17 10:03:07 crc kubenswrapper[4966]: I1217 10:03:07.057224 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fm2ld" event={"ID":"c141723b-23b4-49a3-b743-678314e478e5","Type":"ContainerStarted","Data":"988ae592ad8b1a3c471fefca7335c966dd3f2d07c0c2d0105207a97c853b5489"} Dec 17 10:03:07 crc kubenswrapper[4966]: I1217 10:03:07.082242 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fm2ld" podStartSLOduration=2.613698453 podStartE2EDuration="5.082214239s" podCreationTimestamp="2025-12-17 10:03:02 +0000 UTC" firstStartedPulling="2025-12-17 10:03:04.024415048 +0000 UTC m=+6119.569484990" lastFinishedPulling="2025-12-17 10:03:06.492930844 +0000 UTC m=+6122.038000776" observedRunningTime="2025-12-17 10:03:07.07565966 +0000 UTC m=+6122.620729602" watchObservedRunningTime="2025-12-17 10:03:07.082214239 +0000 UTC m=+6122.627284181" Dec 17 10:03:09 crc kubenswrapper[4966]: I1217 10:03:09.044381 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fvf5s" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="registry-server" probeResult="failure" output=< Dec 17 10:03:09 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:03:09 crc kubenswrapper[4966]: > Dec 17 10:03:12 crc kubenswrapper[4966]: I1217 10:03:12.772735 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:12 crc kubenswrapper[4966]: I1217 10:03:12.773090 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:12 crc kubenswrapper[4966]: I1217 10:03:12.827434 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:13 crc kubenswrapper[4966]: I1217 10:03:13.152460 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:13 crc kubenswrapper[4966]: I1217 10:03:13.203943 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fm2ld"] Dec 17 10:03:15 crc kubenswrapper[4966]: I1217 10:03:15.123167 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fm2ld" podUID="c141723b-23b4-49a3-b743-678314e478e5" containerName="registry-server" containerID="cri-o://988ae592ad8b1a3c471fefca7335c966dd3f2d07c0c2d0105207a97c853b5489" gracePeriod=2 Dec 17 10:03:15 crc kubenswrapper[4966]: I1217 10:03:15.830953 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:03:15 crc kubenswrapper[4966]: E1217 10:03:15.832791 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.137957 4966 generic.go:334] "Generic (PLEG): container finished" podID="c141723b-23b4-49a3-b743-678314e478e5" containerID="988ae592ad8b1a3c471fefca7335c966dd3f2d07c0c2d0105207a97c853b5489" exitCode=0 Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.138067 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fm2ld" event={"ID":"c141723b-23b4-49a3-b743-678314e478e5","Type":"ContainerDied","Data":"988ae592ad8b1a3c471fefca7335c966dd3f2d07c0c2d0105207a97c853b5489"} Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.384745 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.499028 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs995\" (UniqueName: \"kubernetes.io/projected/c141723b-23b4-49a3-b743-678314e478e5-kube-api-access-rs995\") pod \"c141723b-23b4-49a3-b743-678314e478e5\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.499169 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-catalog-content\") pod \"c141723b-23b4-49a3-b743-678314e478e5\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.499202 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-utilities\") pod \"c141723b-23b4-49a3-b743-678314e478e5\" (UID: \"c141723b-23b4-49a3-b743-678314e478e5\") " Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.499832 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-utilities" (OuterVolumeSpecName: "utilities") pod "c141723b-23b4-49a3-b743-678314e478e5" (UID: "c141723b-23b4-49a3-b743-678314e478e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.500154 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.507575 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c141723b-23b4-49a3-b743-678314e478e5-kube-api-access-rs995" (OuterVolumeSpecName: "kube-api-access-rs995") pod "c141723b-23b4-49a3-b743-678314e478e5" (UID: "c141723b-23b4-49a3-b743-678314e478e5"). InnerVolumeSpecName "kube-api-access-rs995". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.515130 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c141723b-23b4-49a3-b743-678314e478e5" (UID: "c141723b-23b4-49a3-b743-678314e478e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.602617 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs995\" (UniqueName: \"kubernetes.io/projected/c141723b-23b4-49a3-b743-678314e478e5-kube-api-access-rs995\") on node \"crc\" DevicePath \"\"" Dec 17 10:03:16 crc kubenswrapper[4966]: I1217 10:03:16.602650 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c141723b-23b4-49a3-b743-678314e478e5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:03:17 crc kubenswrapper[4966]: I1217 10:03:17.149313 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fm2ld" event={"ID":"c141723b-23b4-49a3-b743-678314e478e5","Type":"ContainerDied","Data":"a73a1bfd8efdbecafe90468e871c3655637f6425afed01643d68dcc7ebf9c1ce"} Dec 17 10:03:17 crc kubenswrapper[4966]: I1217 10:03:17.149411 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fm2ld" Dec 17 10:03:17 crc kubenswrapper[4966]: I1217 10:03:17.149591 4966 scope.go:117] "RemoveContainer" containerID="988ae592ad8b1a3c471fefca7335c966dd3f2d07c0c2d0105207a97c853b5489" Dec 17 10:03:17 crc kubenswrapper[4966]: I1217 10:03:17.178195 4966 scope.go:117] "RemoveContainer" containerID="71d6fe931446bcb2ecec99f85415841efb1e65fcc7391b4960313e5d9df5dd7c" Dec 17 10:03:17 crc kubenswrapper[4966]: I1217 10:03:17.178347 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fm2ld"] Dec 17 10:03:17 crc kubenswrapper[4966]: I1217 10:03:17.199527 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fm2ld"] Dec 17 10:03:17 crc kubenswrapper[4966]: I1217 10:03:17.203190 4966 scope.go:117] "RemoveContainer" containerID="7686d79bd9e6ff9e7d53a0a80a37a1cdc05d6b4f80a314b6975d201f3a0daf5a" Dec 17 10:03:18 crc kubenswrapper[4966]: I1217 10:03:18.044336 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:03:18 crc kubenswrapper[4966]: I1217 10:03:18.096007 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:03:18 crc kubenswrapper[4966]: I1217 10:03:18.843814 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c141723b-23b4-49a3-b743-678314e478e5" path="/var/lib/kubelet/pods/c141723b-23b4-49a3-b743-678314e478e5/volumes" Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.028260 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fvf5s"] Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.165888 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fvf5s" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="registry-server" containerID="cri-o://36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff" gracePeriod=2 Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.690031 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.777581 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-utilities\") pod \"97effbb7-ea20-49ec-a328-182815c3c087\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.777886 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-catalog-content\") pod \"97effbb7-ea20-49ec-a328-182815c3c087\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.778104 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztjkn\" (UniqueName: \"kubernetes.io/projected/97effbb7-ea20-49ec-a328-182815c3c087-kube-api-access-ztjkn\") pod \"97effbb7-ea20-49ec-a328-182815c3c087\" (UID: \"97effbb7-ea20-49ec-a328-182815c3c087\") " Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.778195 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-utilities" (OuterVolumeSpecName: "utilities") pod "97effbb7-ea20-49ec-a328-182815c3c087" (UID: "97effbb7-ea20-49ec-a328-182815c3c087"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.778662 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.784517 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97effbb7-ea20-49ec-a328-182815c3c087-kube-api-access-ztjkn" (OuterVolumeSpecName: "kube-api-access-ztjkn") pod "97effbb7-ea20-49ec-a328-182815c3c087" (UID: "97effbb7-ea20-49ec-a328-182815c3c087"). InnerVolumeSpecName "kube-api-access-ztjkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.880689 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztjkn\" (UniqueName: \"kubernetes.io/projected/97effbb7-ea20-49ec-a328-182815c3c087-kube-api-access-ztjkn\") on node \"crc\" DevicePath \"\"" Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.883426 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97effbb7-ea20-49ec-a328-182815c3c087" (UID: "97effbb7-ea20-49ec-a328-182815c3c087"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:03:19 crc kubenswrapper[4966]: I1217 10:03:19.983896 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97effbb7-ea20-49ec-a328-182815c3c087-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.174495 4966 generic.go:334] "Generic (PLEG): container finished" podID="97effbb7-ea20-49ec-a328-182815c3c087" containerID="36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff" exitCode=0 Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.174533 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvf5s" event={"ID":"97effbb7-ea20-49ec-a328-182815c3c087","Type":"ContainerDied","Data":"36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff"} Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.174558 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fvf5s" event={"ID":"97effbb7-ea20-49ec-a328-182815c3c087","Type":"ContainerDied","Data":"d87c0b9550357885eacaaedf7a8bdb2f615818af83adb684dc245d46a794c92a"} Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.174576 4966 scope.go:117] "RemoveContainer" containerID="36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.174641 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fvf5s" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.198053 4966 scope.go:117] "RemoveContainer" containerID="d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.219622 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fvf5s"] Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.229346 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fvf5s"] Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.231142 4966 scope.go:117] "RemoveContainer" containerID="de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.281529 4966 scope.go:117] "RemoveContainer" containerID="36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff" Dec 17 10:03:20 crc kubenswrapper[4966]: E1217 10:03:20.284339 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff\": container with ID starting with 36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff not found: ID does not exist" containerID="36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.284376 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff"} err="failed to get container status \"36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff\": rpc error: code = NotFound desc = could not find container \"36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff\": container with ID starting with 36d487b25de8b21703ae0a69fcd26bb40f7f82e3e4edf644382b4e0375d33cff not found: ID does not exist" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.284398 4966 scope.go:117] "RemoveContainer" containerID="d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646" Dec 17 10:03:20 crc kubenswrapper[4966]: E1217 10:03:20.284841 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646\": container with ID starting with d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646 not found: ID does not exist" containerID="d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.284865 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646"} err="failed to get container status \"d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646\": rpc error: code = NotFound desc = could not find container \"d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646\": container with ID starting with d61505460ad1904a29c83fadd266adb44c4ec217e1df5016fc6a1473ee6ca646 not found: ID does not exist" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.284911 4966 scope.go:117] "RemoveContainer" containerID="de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8" Dec 17 10:03:20 crc kubenswrapper[4966]: E1217 10:03:20.285169 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8\": container with ID starting with de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8 not found: ID does not exist" containerID="de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.285185 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8"} err="failed to get container status \"de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8\": rpc error: code = NotFound desc = could not find container \"de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8\": container with ID starting with de5549d8686a3999da2a45c4707d1a8bee6949698d6ac3eaad791c8fdccfdbf8 not found: ID does not exist" Dec 17 10:03:20 crc kubenswrapper[4966]: I1217 10:03:20.843231 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97effbb7-ea20-49ec-a328-182815c3c087" path="/var/lib/kubelet/pods/97effbb7-ea20-49ec-a328-182815c3c087/volumes" Dec 17 10:03:30 crc kubenswrapper[4966]: I1217 10:03:30.830446 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:03:30 crc kubenswrapper[4966]: E1217 10:03:30.832461 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:03:44 crc kubenswrapper[4966]: I1217 10:03:44.839685 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:03:44 crc kubenswrapper[4966]: E1217 10:03:44.840591 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:03:55 crc kubenswrapper[4966]: I1217 10:03:55.830671 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:03:55 crc kubenswrapper[4966]: E1217 10:03:55.831597 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:04:10 crc kubenswrapper[4966]: I1217 10:04:10.830678 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:04:10 crc kubenswrapper[4966]: E1217 10:04:10.831648 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:04:23 crc kubenswrapper[4966]: I1217 10:04:23.830135 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:04:23 crc kubenswrapper[4966]: E1217 10:04:23.830880 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:04:37 crc kubenswrapper[4966]: I1217 10:04:37.831153 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:04:37 crc kubenswrapper[4966]: E1217 10:04:37.832268 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:04:52 crc kubenswrapper[4966]: I1217 10:04:52.831793 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:04:52 crc kubenswrapper[4966]: E1217 10:04:52.832811 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:05:03 crc kubenswrapper[4966]: I1217 10:05:03.831085 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:05:03 crc kubenswrapper[4966]: E1217 10:05:03.831597 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:05:17 crc kubenswrapper[4966]: I1217 10:05:17.830722 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:05:17 crc kubenswrapper[4966]: E1217 10:05:17.831497 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.657179 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nrwrr"] Dec 17 10:05:19 crc kubenswrapper[4966]: E1217 10:05:19.658154 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="registry-server" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.658171 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="registry-server" Dec 17 10:05:19 crc kubenswrapper[4966]: E1217 10:05:19.658183 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="extract-content" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.658189 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="extract-content" Dec 17 10:05:19 crc kubenswrapper[4966]: E1217 10:05:19.658210 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c141723b-23b4-49a3-b743-678314e478e5" containerName="extract-utilities" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.658216 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c141723b-23b4-49a3-b743-678314e478e5" containerName="extract-utilities" Dec 17 10:05:19 crc kubenswrapper[4966]: E1217 10:05:19.658233 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c141723b-23b4-49a3-b743-678314e478e5" containerName="extract-content" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.658239 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c141723b-23b4-49a3-b743-678314e478e5" containerName="extract-content" Dec 17 10:05:19 crc kubenswrapper[4966]: E1217 10:05:19.658251 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c141723b-23b4-49a3-b743-678314e478e5" containerName="registry-server" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.658258 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c141723b-23b4-49a3-b743-678314e478e5" containerName="registry-server" Dec 17 10:05:19 crc kubenswrapper[4966]: E1217 10:05:19.658272 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="extract-utilities" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.658279 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="extract-utilities" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.658508 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c141723b-23b4-49a3-b743-678314e478e5" containerName="registry-server" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.658520 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="97effbb7-ea20-49ec-a328-182815c3c087" containerName="registry-server" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.660127 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.667208 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrwrr"] Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.782134 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsx5j\" (UniqueName: \"kubernetes.io/projected/da967833-d2bf-4844-ab47-049bd10bab86-kube-api-access-tsx5j\") pod \"certified-operators-nrwrr\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.782342 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-catalog-content\") pod \"certified-operators-nrwrr\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.782414 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-utilities\") pod \"certified-operators-nrwrr\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.883967 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-catalog-content\") pod \"certified-operators-nrwrr\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.884076 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-utilities\") pod \"certified-operators-nrwrr\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.884197 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsx5j\" (UniqueName: \"kubernetes.io/projected/da967833-d2bf-4844-ab47-049bd10bab86-kube-api-access-tsx5j\") pod \"certified-operators-nrwrr\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.884518 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-catalog-content\") pod \"certified-operators-nrwrr\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.884731 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-utilities\") pod \"certified-operators-nrwrr\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.919849 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsx5j\" (UniqueName: \"kubernetes.io/projected/da967833-d2bf-4844-ab47-049bd10bab86-kube-api-access-tsx5j\") pod \"certified-operators-nrwrr\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:19 crc kubenswrapper[4966]: I1217 10:05:19.990166 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:20 crc kubenswrapper[4966]: I1217 10:05:20.582695 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrwrr"] Dec 17 10:05:21 crc kubenswrapper[4966]: I1217 10:05:21.304620 4966 generic.go:334] "Generic (PLEG): container finished" podID="da967833-d2bf-4844-ab47-049bd10bab86" containerID="69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592" exitCode=0 Dec 17 10:05:21 crc kubenswrapper[4966]: I1217 10:05:21.304661 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrwrr" event={"ID":"da967833-d2bf-4844-ab47-049bd10bab86","Type":"ContainerDied","Data":"69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592"} Dec 17 10:05:21 crc kubenswrapper[4966]: I1217 10:05:21.304958 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrwrr" event={"ID":"da967833-d2bf-4844-ab47-049bd10bab86","Type":"ContainerStarted","Data":"81b42cd19f9dcfc038f02d6e29d95765f2a7fefdba03b1c32e04d124b3d59f9e"} Dec 17 10:05:23 crc kubenswrapper[4966]: I1217 10:05:23.333911 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrwrr" event={"ID":"da967833-d2bf-4844-ab47-049bd10bab86","Type":"ContainerStarted","Data":"1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7"} Dec 17 10:05:24 crc kubenswrapper[4966]: I1217 10:05:24.347798 4966 generic.go:334] "Generic (PLEG): container finished" podID="da967833-d2bf-4844-ab47-049bd10bab86" containerID="1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7" exitCode=0 Dec 17 10:05:24 crc kubenswrapper[4966]: I1217 10:05:24.347958 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrwrr" event={"ID":"da967833-d2bf-4844-ab47-049bd10bab86","Type":"ContainerDied","Data":"1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7"} Dec 17 10:05:25 crc kubenswrapper[4966]: I1217 10:05:25.359419 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrwrr" event={"ID":"da967833-d2bf-4844-ab47-049bd10bab86","Type":"ContainerStarted","Data":"a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e"} Dec 17 10:05:25 crc kubenswrapper[4966]: I1217 10:05:25.383104 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nrwrr" podStartSLOduration=2.735313804 podStartE2EDuration="6.383084367s" podCreationTimestamp="2025-12-17 10:05:19 +0000 UTC" firstStartedPulling="2025-12-17 10:05:21.306841186 +0000 UTC m=+6256.851911128" lastFinishedPulling="2025-12-17 10:05:24.954611749 +0000 UTC m=+6260.499681691" observedRunningTime="2025-12-17 10:05:25.380111426 +0000 UTC m=+6260.925181368" watchObservedRunningTime="2025-12-17 10:05:25.383084367 +0000 UTC m=+6260.928154309" Dec 17 10:05:29 crc kubenswrapper[4966]: I1217 10:05:29.991080 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:29 crc kubenswrapper[4966]: I1217 10:05:29.991652 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:30 crc kubenswrapper[4966]: I1217 10:05:30.040776 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:30 crc kubenswrapper[4966]: I1217 10:05:30.465166 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:30 crc kubenswrapper[4966]: I1217 10:05:30.516545 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrwrr"] Dec 17 10:05:30 crc kubenswrapper[4966]: I1217 10:05:30.830763 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:05:30 crc kubenswrapper[4966]: E1217 10:05:30.831401 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:05:32 crc kubenswrapper[4966]: I1217 10:05:32.425409 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nrwrr" podUID="da967833-d2bf-4844-ab47-049bd10bab86" containerName="registry-server" containerID="cri-o://a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e" gracePeriod=2 Dec 17 10:05:32 crc kubenswrapper[4966]: I1217 10:05:32.921766 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.087208 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsx5j\" (UniqueName: \"kubernetes.io/projected/da967833-d2bf-4844-ab47-049bd10bab86-kube-api-access-tsx5j\") pod \"da967833-d2bf-4844-ab47-049bd10bab86\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.087375 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-utilities\") pod \"da967833-d2bf-4844-ab47-049bd10bab86\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.087516 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-catalog-content\") pod \"da967833-d2bf-4844-ab47-049bd10bab86\" (UID: \"da967833-d2bf-4844-ab47-049bd10bab86\") " Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.089123 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-utilities" (OuterVolumeSpecName: "utilities") pod "da967833-d2bf-4844-ab47-049bd10bab86" (UID: "da967833-d2bf-4844-ab47-049bd10bab86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.096565 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da967833-d2bf-4844-ab47-049bd10bab86-kube-api-access-tsx5j" (OuterVolumeSpecName: "kube-api-access-tsx5j") pod "da967833-d2bf-4844-ab47-049bd10bab86" (UID: "da967833-d2bf-4844-ab47-049bd10bab86"). InnerVolumeSpecName "kube-api-access-tsx5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.146486 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da967833-d2bf-4844-ab47-049bd10bab86" (UID: "da967833-d2bf-4844-ab47-049bd10bab86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.189461 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.189493 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsx5j\" (UniqueName: \"kubernetes.io/projected/da967833-d2bf-4844-ab47-049bd10bab86-kube-api-access-tsx5j\") on node \"crc\" DevicePath \"\"" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.189502 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da967833-d2bf-4844-ab47-049bd10bab86-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.438011 4966 generic.go:334] "Generic (PLEG): container finished" podID="da967833-d2bf-4844-ab47-049bd10bab86" containerID="a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e" exitCode=0 Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.438068 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrwrr" event={"ID":"da967833-d2bf-4844-ab47-049bd10bab86","Type":"ContainerDied","Data":"a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e"} Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.438107 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrwrr" event={"ID":"da967833-d2bf-4844-ab47-049bd10bab86","Type":"ContainerDied","Data":"81b42cd19f9dcfc038f02d6e29d95765f2a7fefdba03b1c32e04d124b3d59f9e"} Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.438130 4966 scope.go:117] "RemoveContainer" containerID="a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.438329 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrwrr" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.495647 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrwrr"] Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.506558 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nrwrr"] Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.515064 4966 scope.go:117] "RemoveContainer" containerID="1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.540630 4966 scope.go:117] "RemoveContainer" containerID="69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.602436 4966 scope.go:117] "RemoveContainer" containerID="a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e" Dec 17 10:05:33 crc kubenswrapper[4966]: E1217 10:05:33.603021 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e\": container with ID starting with a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e not found: ID does not exist" containerID="a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.603193 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e"} err="failed to get container status \"a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e\": rpc error: code = NotFound desc = could not find container \"a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e\": container with ID starting with a0216e0858879697559ebd9c0c7524787ead46cc1c0d624ac3f65e3c6a9c750e not found: ID does not exist" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.603231 4966 scope.go:117] "RemoveContainer" containerID="1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7" Dec 17 10:05:33 crc kubenswrapper[4966]: E1217 10:05:33.603594 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7\": container with ID starting with 1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7 not found: ID does not exist" containerID="1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.603640 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7"} err="failed to get container status \"1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7\": rpc error: code = NotFound desc = could not find container \"1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7\": container with ID starting with 1d43ea641fb446f188c8bcad3754efad16cdb01d71e4a65efc21c1350b1acab7 not found: ID does not exist" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.603656 4966 scope.go:117] "RemoveContainer" containerID="69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592" Dec 17 10:05:33 crc kubenswrapper[4966]: E1217 10:05:33.604366 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592\": container with ID starting with 69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592 not found: ID does not exist" containerID="69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592" Dec 17 10:05:33 crc kubenswrapper[4966]: I1217 10:05:33.604421 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592"} err="failed to get container status \"69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592\": rpc error: code = NotFound desc = could not find container \"69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592\": container with ID starting with 69a1234407d5c2bdc1da4363f811b9d269a9d3f2a9f4583a8791a7c355781592 not found: ID does not exist" Dec 17 10:05:34 crc kubenswrapper[4966]: I1217 10:05:34.844317 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da967833-d2bf-4844-ab47-049bd10bab86" path="/var/lib/kubelet/pods/da967833-d2bf-4844-ab47-049bd10bab86/volumes" Dec 17 10:05:43 crc kubenswrapper[4966]: I1217 10:05:43.831072 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:05:43 crc kubenswrapper[4966]: E1217 10:05:43.831934 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.038031 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j57m6"] Dec 17 10:05:46 crc kubenswrapper[4966]: E1217 10:05:46.039187 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da967833-d2bf-4844-ab47-049bd10bab86" containerName="registry-server" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.039203 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="da967833-d2bf-4844-ab47-049bd10bab86" containerName="registry-server" Dec 17 10:05:46 crc kubenswrapper[4966]: E1217 10:05:46.039223 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da967833-d2bf-4844-ab47-049bd10bab86" containerName="extract-utilities" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.039232 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="da967833-d2bf-4844-ab47-049bd10bab86" containerName="extract-utilities" Dec 17 10:05:46 crc kubenswrapper[4966]: E1217 10:05:46.039259 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da967833-d2bf-4844-ab47-049bd10bab86" containerName="extract-content" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.039267 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="da967833-d2bf-4844-ab47-049bd10bab86" containerName="extract-content" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.039529 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="da967833-d2bf-4844-ab47-049bd10bab86" containerName="registry-server" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.041643 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.043318 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jbrb\" (UniqueName: \"kubernetes.io/projected/3ba50c48-ff25-4da2-9233-d53639c4d997-kube-api-access-8jbrb\") pod \"community-operators-j57m6\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.043444 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-catalog-content\") pod \"community-operators-j57m6\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.043513 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-utilities\") pod \"community-operators-j57m6\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.065264 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j57m6"] Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.153821 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jbrb\" (UniqueName: \"kubernetes.io/projected/3ba50c48-ff25-4da2-9233-d53639c4d997-kube-api-access-8jbrb\") pod \"community-operators-j57m6\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.154068 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-catalog-content\") pod \"community-operators-j57m6\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.154168 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-utilities\") pod \"community-operators-j57m6\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.154909 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-utilities\") pod \"community-operators-j57m6\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.156165 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-catalog-content\") pod \"community-operators-j57m6\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.208185 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jbrb\" (UniqueName: \"kubernetes.io/projected/3ba50c48-ff25-4da2-9233-d53639c4d997-kube-api-access-8jbrb\") pod \"community-operators-j57m6\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.369021 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:46 crc kubenswrapper[4966]: I1217 10:05:46.892822 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j57m6"] Dec 17 10:05:47 crc kubenswrapper[4966]: I1217 10:05:47.589829 4966 generic.go:334] "Generic (PLEG): container finished" podID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerID="189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9" exitCode=0 Dec 17 10:05:47 crc kubenswrapper[4966]: I1217 10:05:47.589983 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57m6" event={"ID":"3ba50c48-ff25-4da2-9233-d53639c4d997","Type":"ContainerDied","Data":"189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9"} Dec 17 10:05:47 crc kubenswrapper[4966]: I1217 10:05:47.590156 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57m6" event={"ID":"3ba50c48-ff25-4da2-9233-d53639c4d997","Type":"ContainerStarted","Data":"19c4082156d60a0d0cf70a509a328f2fe46afa4908e84ea630d7d6a75dc65798"} Dec 17 10:05:49 crc kubenswrapper[4966]: I1217 10:05:49.628967 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57m6" event={"ID":"3ba50c48-ff25-4da2-9233-d53639c4d997","Type":"ContainerStarted","Data":"257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95"} Dec 17 10:05:50 crc kubenswrapper[4966]: I1217 10:05:50.640373 4966 generic.go:334] "Generic (PLEG): container finished" podID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerID="257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95" exitCode=0 Dec 17 10:05:50 crc kubenswrapper[4966]: I1217 10:05:50.640454 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57m6" event={"ID":"3ba50c48-ff25-4da2-9233-d53639c4d997","Type":"ContainerDied","Data":"257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95"} Dec 17 10:05:51 crc kubenswrapper[4966]: I1217 10:05:51.658092 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57m6" event={"ID":"3ba50c48-ff25-4da2-9233-d53639c4d997","Type":"ContainerStarted","Data":"7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478"} Dec 17 10:05:51 crc kubenswrapper[4966]: I1217 10:05:51.685798 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j57m6" podStartSLOduration=2.124886526 podStartE2EDuration="5.685783059s" podCreationTimestamp="2025-12-17 10:05:46 +0000 UTC" firstStartedPulling="2025-12-17 10:05:47.592841563 +0000 UTC m=+6283.137911505" lastFinishedPulling="2025-12-17 10:05:51.153738096 +0000 UTC m=+6286.698808038" observedRunningTime="2025-12-17 10:05:51.684619917 +0000 UTC m=+6287.229689859" watchObservedRunningTime="2025-12-17 10:05:51.685783059 +0000 UTC m=+6287.230853001" Dec 17 10:05:56 crc kubenswrapper[4966]: I1217 10:05:56.370000 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:56 crc kubenswrapper[4966]: I1217 10:05:56.370619 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:56 crc kubenswrapper[4966]: I1217 10:05:56.420635 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:56 crc kubenswrapper[4966]: I1217 10:05:56.757771 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:56 crc kubenswrapper[4966]: I1217 10:05:56.819112 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j57m6"] Dec 17 10:05:57 crc kubenswrapper[4966]: I1217 10:05:57.830528 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:05:57 crc kubenswrapper[4966]: E1217 10:05:57.830764 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:05:58 crc kubenswrapper[4966]: I1217 10:05:58.725969 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j57m6" podUID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerName="registry-server" containerID="cri-o://7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478" gracePeriod=2 Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.238545 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.342969 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-utilities\") pod \"3ba50c48-ff25-4da2-9233-d53639c4d997\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.343242 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jbrb\" (UniqueName: \"kubernetes.io/projected/3ba50c48-ff25-4da2-9233-d53639c4d997-kube-api-access-8jbrb\") pod \"3ba50c48-ff25-4da2-9233-d53639c4d997\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.343366 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-catalog-content\") pod \"3ba50c48-ff25-4da2-9233-d53639c4d997\" (UID: \"3ba50c48-ff25-4da2-9233-d53639c4d997\") " Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.345382 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-utilities" (OuterVolumeSpecName: "utilities") pod "3ba50c48-ff25-4da2-9233-d53639c4d997" (UID: "3ba50c48-ff25-4da2-9233-d53639c4d997"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.352280 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ba50c48-ff25-4da2-9233-d53639c4d997-kube-api-access-8jbrb" (OuterVolumeSpecName: "kube-api-access-8jbrb") pod "3ba50c48-ff25-4da2-9233-d53639c4d997" (UID: "3ba50c48-ff25-4da2-9233-d53639c4d997"). InnerVolumeSpecName "kube-api-access-8jbrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.406746 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ba50c48-ff25-4da2-9233-d53639c4d997" (UID: "3ba50c48-ff25-4da2-9233-d53639c4d997"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.446234 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jbrb\" (UniqueName: \"kubernetes.io/projected/3ba50c48-ff25-4da2-9233-d53639c4d997-kube-api-access-8jbrb\") on node \"crc\" DevicePath \"\"" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.446514 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.446609 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ba50c48-ff25-4da2-9233-d53639c4d997-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.736681 4966 generic.go:334] "Generic (PLEG): container finished" podID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerID="7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478" exitCode=0 Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.736723 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57m6" event={"ID":"3ba50c48-ff25-4da2-9233-d53639c4d997","Type":"ContainerDied","Data":"7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478"} Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.736748 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j57m6" event={"ID":"3ba50c48-ff25-4da2-9233-d53639c4d997","Type":"ContainerDied","Data":"19c4082156d60a0d0cf70a509a328f2fe46afa4908e84ea630d7d6a75dc65798"} Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.736765 4966 scope.go:117] "RemoveContainer" containerID="7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.736933 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j57m6" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.761732 4966 scope.go:117] "RemoveContainer" containerID="257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.784974 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j57m6"] Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.795774 4966 scope.go:117] "RemoveContainer" containerID="189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.799027 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j57m6"] Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.839268 4966 scope.go:117] "RemoveContainer" containerID="7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478" Dec 17 10:05:59 crc kubenswrapper[4966]: E1217 10:05:59.839690 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478\": container with ID starting with 7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478 not found: ID does not exist" containerID="7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.839740 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478"} err="failed to get container status \"7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478\": rpc error: code = NotFound desc = could not find container \"7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478\": container with ID starting with 7cb23ce38b3a1c110350bf6c24eb7438c1741a16b0cde460cddd0f1aece9e478 not found: ID does not exist" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.839765 4966 scope.go:117] "RemoveContainer" containerID="257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95" Dec 17 10:05:59 crc kubenswrapper[4966]: E1217 10:05:59.840066 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95\": container with ID starting with 257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95 not found: ID does not exist" containerID="257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.840092 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95"} err="failed to get container status \"257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95\": rpc error: code = NotFound desc = could not find container \"257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95\": container with ID starting with 257ab2a696af6d483143b964ee91f031978250f80da93dbf4121d93a9a97be95 not found: ID does not exist" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.840106 4966 scope.go:117] "RemoveContainer" containerID="189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9" Dec 17 10:05:59 crc kubenswrapper[4966]: E1217 10:05:59.841102 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9\": container with ID starting with 189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9 not found: ID does not exist" containerID="189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9" Dec 17 10:05:59 crc kubenswrapper[4966]: I1217 10:05:59.841126 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9"} err="failed to get container status \"189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9\": rpc error: code = NotFound desc = could not find container \"189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9\": container with ID starting with 189656af2d65369d17c6692e5f7880f254e721affbe951987c32b5e87f5c31f9 not found: ID does not exist" Dec 17 10:06:00 crc kubenswrapper[4966]: I1217 10:06:00.841409 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ba50c48-ff25-4da2-9233-d53639c4d997" path="/var/lib/kubelet/pods/3ba50c48-ff25-4da2-9233-d53639c4d997/volumes" Dec 17 10:06:11 crc kubenswrapper[4966]: I1217 10:06:11.831379 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:06:11 crc kubenswrapper[4966]: E1217 10:06:11.832336 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:06:24 crc kubenswrapper[4966]: I1217 10:06:24.837113 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:06:26 crc kubenswrapper[4966]: I1217 10:06:26.042345 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"252200fafb3d0845e253b308a0cbdce98a9ba720a50bf3096520c974add44fb9"} Dec 17 10:08:46 crc kubenswrapper[4966]: I1217 10:08:46.807653 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:08:46 crc kubenswrapper[4966]: I1217 10:08:46.808448 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:09:16 crc kubenswrapper[4966]: I1217 10:09:16.807575 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:09:16 crc kubenswrapper[4966]: I1217 10:09:16.808178 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:09:46 crc kubenswrapper[4966]: I1217 10:09:46.807533 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:09:46 crc kubenswrapper[4966]: I1217 10:09:46.808126 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:09:46 crc kubenswrapper[4966]: I1217 10:09:46.808203 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:09:46 crc kubenswrapper[4966]: I1217 10:09:46.809359 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"252200fafb3d0845e253b308a0cbdce98a9ba720a50bf3096520c974add44fb9"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:09:46 crc kubenswrapper[4966]: I1217 10:09:46.809473 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://252200fafb3d0845e253b308a0cbdce98a9ba720a50bf3096520c974add44fb9" gracePeriod=600 Dec 17 10:09:47 crc kubenswrapper[4966]: I1217 10:09:47.844042 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="252200fafb3d0845e253b308a0cbdce98a9ba720a50bf3096520c974add44fb9" exitCode=0 Dec 17 10:09:47 crc kubenswrapper[4966]: I1217 10:09:47.844097 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"252200fafb3d0845e253b308a0cbdce98a9ba720a50bf3096520c974add44fb9"} Dec 17 10:09:47 crc kubenswrapper[4966]: I1217 10:09:47.844533 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2"} Dec 17 10:09:47 crc kubenswrapper[4966]: I1217 10:09:47.844591 4966 scope.go:117] "RemoveContainer" containerID="3925fe0f85697a44b78066592b5705d12ffc6949cbb6c7339809c0b39b52e0ad" Dec 17 10:12:16 crc kubenswrapper[4966]: I1217 10:12:16.807517 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:12:16 crc kubenswrapper[4966]: I1217 10:12:16.808333 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:12:46 crc kubenswrapper[4966]: I1217 10:12:46.807380 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:12:46 crc kubenswrapper[4966]: I1217 10:12:46.808632 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:13:16 crc kubenswrapper[4966]: I1217 10:13:16.807436 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:13:16 crc kubenswrapper[4966]: I1217 10:13:16.808075 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:13:16 crc kubenswrapper[4966]: I1217 10:13:16.808163 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:13:16 crc kubenswrapper[4966]: I1217 10:13:16.809224 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:13:16 crc kubenswrapper[4966]: I1217 10:13:16.809301 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" gracePeriod=600 Dec 17 10:13:17 crc kubenswrapper[4966]: E1217 10:13:17.005412 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:13:17 crc kubenswrapper[4966]: I1217 10:13:17.536971 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" exitCode=0 Dec 17 10:13:17 crc kubenswrapper[4966]: I1217 10:13:17.537013 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2"} Dec 17 10:13:17 crc kubenswrapper[4966]: I1217 10:13:17.537062 4966 scope.go:117] "RemoveContainer" containerID="252200fafb3d0845e253b308a0cbdce98a9ba720a50bf3096520c974add44fb9" Dec 17 10:13:17 crc kubenswrapper[4966]: I1217 10:13:17.537718 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:13:17 crc kubenswrapper[4966]: E1217 10:13:17.538085 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:13:30 crc kubenswrapper[4966]: I1217 10:13:30.831017 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:13:30 crc kubenswrapper[4966]: E1217 10:13:30.832184 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.371259 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qvd85"] Dec 17 10:13:31 crc kubenswrapper[4966]: E1217 10:13:31.371714 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerName="extract-content" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.371738 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerName="extract-content" Dec 17 10:13:31 crc kubenswrapper[4966]: E1217 10:13:31.371753 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerName="extract-utilities" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.371761 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerName="extract-utilities" Dec 17 10:13:31 crc kubenswrapper[4966]: E1217 10:13:31.371809 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerName="registry-server" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.371816 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerName="registry-server" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.372021 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ba50c48-ff25-4da2-9233-d53639c4d997" containerName="registry-server" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.374382 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.393658 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvd85"] Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.439733 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zfhl\" (UniqueName: \"kubernetes.io/projected/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-kube-api-access-6zfhl\") pod \"redhat-marketplace-qvd85\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.439935 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-catalog-content\") pod \"redhat-marketplace-qvd85\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.439968 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-utilities\") pod \"redhat-marketplace-qvd85\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.541593 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zfhl\" (UniqueName: \"kubernetes.io/projected/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-kube-api-access-6zfhl\") pod \"redhat-marketplace-qvd85\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.541717 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-catalog-content\") pod \"redhat-marketplace-qvd85\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.541741 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-utilities\") pod \"redhat-marketplace-qvd85\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.542656 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-catalog-content\") pod \"redhat-marketplace-qvd85\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.542710 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-utilities\") pod \"redhat-marketplace-qvd85\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.569368 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hj8hv"] Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.572694 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.574180 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zfhl\" (UniqueName: \"kubernetes.io/projected/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-kube-api-access-6zfhl\") pod \"redhat-marketplace-qvd85\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.582308 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hj8hv"] Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.645477 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfhrx\" (UniqueName: \"kubernetes.io/projected/826113b3-9fc8-41c6-948f-4251cc7c430d-kube-api-access-pfhrx\") pod \"redhat-operators-hj8hv\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.645592 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-utilities\") pod \"redhat-operators-hj8hv\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.645644 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-catalog-content\") pod \"redhat-operators-hj8hv\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.695622 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.748252 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfhrx\" (UniqueName: \"kubernetes.io/projected/826113b3-9fc8-41c6-948f-4251cc7c430d-kube-api-access-pfhrx\") pod \"redhat-operators-hj8hv\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.748454 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-utilities\") pod \"redhat-operators-hj8hv\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.748519 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-catalog-content\") pod \"redhat-operators-hj8hv\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.749134 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-catalog-content\") pod \"redhat-operators-hj8hv\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.749266 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-utilities\") pod \"redhat-operators-hj8hv\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.771074 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfhrx\" (UniqueName: \"kubernetes.io/projected/826113b3-9fc8-41c6-948f-4251cc7c430d-kube-api-access-pfhrx\") pod \"redhat-operators-hj8hv\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:31 crc kubenswrapper[4966]: I1217 10:13:31.944497 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:32 crc kubenswrapper[4966]: I1217 10:13:32.423365 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvd85"] Dec 17 10:13:32 crc kubenswrapper[4966]: I1217 10:13:32.565031 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hj8hv"] Dec 17 10:13:32 crc kubenswrapper[4966]: I1217 10:13:32.692108 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8hv" event={"ID":"826113b3-9fc8-41c6-948f-4251cc7c430d","Type":"ContainerStarted","Data":"e4e6e7b7cc5c2e3ff6cebbf4406d5b3a44d848e88475eb28d2545da090cddfd7"} Dec 17 10:13:32 crc kubenswrapper[4966]: I1217 10:13:32.694728 4966 generic.go:334] "Generic (PLEG): container finished" podID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerID="4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8" exitCode=0 Dec 17 10:13:32 crc kubenswrapper[4966]: I1217 10:13:32.694786 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvd85" event={"ID":"3e4d1c1d-8170-4d42-b9d3-7128e64766c9","Type":"ContainerDied","Data":"4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8"} Dec 17 10:13:32 crc kubenswrapper[4966]: I1217 10:13:32.694809 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvd85" event={"ID":"3e4d1c1d-8170-4d42-b9d3-7128e64766c9","Type":"ContainerStarted","Data":"cb381ab59ebfc98a18eee5765f4585c5357d8dcdb4eef7e27e1aced90208b2a8"} Dec 17 10:13:32 crc kubenswrapper[4966]: I1217 10:13:32.697097 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 10:13:33 crc kubenswrapper[4966]: I1217 10:13:33.704838 4966 generic.go:334] "Generic (PLEG): container finished" podID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerID="943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb" exitCode=0 Dec 17 10:13:33 crc kubenswrapper[4966]: I1217 10:13:33.705156 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8hv" event={"ID":"826113b3-9fc8-41c6-948f-4251cc7c430d","Type":"ContainerDied","Data":"943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb"} Dec 17 10:13:34 crc kubenswrapper[4966]: I1217 10:13:34.718505 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvd85" event={"ID":"3e4d1c1d-8170-4d42-b9d3-7128e64766c9","Type":"ContainerStarted","Data":"6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c"} Dec 17 10:13:35 crc kubenswrapper[4966]: I1217 10:13:35.728460 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8hv" event={"ID":"826113b3-9fc8-41c6-948f-4251cc7c430d","Type":"ContainerStarted","Data":"003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334"} Dec 17 10:13:35 crc kubenswrapper[4966]: I1217 10:13:35.735196 4966 generic.go:334] "Generic (PLEG): container finished" podID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerID="6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c" exitCode=0 Dec 17 10:13:35 crc kubenswrapper[4966]: I1217 10:13:35.735255 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvd85" event={"ID":"3e4d1c1d-8170-4d42-b9d3-7128e64766c9","Type":"ContainerDied","Data":"6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c"} Dec 17 10:13:37 crc kubenswrapper[4966]: I1217 10:13:37.753996 4966 generic.go:334] "Generic (PLEG): container finished" podID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerID="003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334" exitCode=0 Dec 17 10:13:37 crc kubenswrapper[4966]: I1217 10:13:37.754103 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8hv" event={"ID":"826113b3-9fc8-41c6-948f-4251cc7c430d","Type":"ContainerDied","Data":"003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334"} Dec 17 10:13:37 crc kubenswrapper[4966]: I1217 10:13:37.757230 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvd85" event={"ID":"3e4d1c1d-8170-4d42-b9d3-7128e64766c9","Type":"ContainerStarted","Data":"50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26"} Dec 17 10:13:37 crc kubenswrapper[4966]: I1217 10:13:37.807189 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qvd85" podStartSLOduration=3.027320156 podStartE2EDuration="6.806979308s" podCreationTimestamp="2025-12-17 10:13:31 +0000 UTC" firstStartedPulling="2025-12-17 10:13:32.696489614 +0000 UTC m=+6748.241559556" lastFinishedPulling="2025-12-17 10:13:36.476148766 +0000 UTC m=+6752.021218708" observedRunningTime="2025-12-17 10:13:37.803423061 +0000 UTC m=+6753.348493003" watchObservedRunningTime="2025-12-17 10:13:37.806979308 +0000 UTC m=+6753.352049250" Dec 17 10:13:38 crc kubenswrapper[4966]: I1217 10:13:38.769641 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8hv" event={"ID":"826113b3-9fc8-41c6-948f-4251cc7c430d","Type":"ContainerStarted","Data":"0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a"} Dec 17 10:13:38 crc kubenswrapper[4966]: I1217 10:13:38.796548 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hj8hv" podStartSLOduration=3.3230007329999998 podStartE2EDuration="7.796528951s" podCreationTimestamp="2025-12-17 10:13:31 +0000 UTC" firstStartedPulling="2025-12-17 10:13:33.715805919 +0000 UTC m=+6749.260875861" lastFinishedPulling="2025-12-17 10:13:38.189334127 +0000 UTC m=+6753.734404079" observedRunningTime="2025-12-17 10:13:38.794075233 +0000 UTC m=+6754.339145195" watchObservedRunningTime="2025-12-17 10:13:38.796528951 +0000 UTC m=+6754.341598893" Dec 17 10:13:41 crc kubenswrapper[4966]: I1217 10:13:41.696169 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:41 crc kubenswrapper[4966]: I1217 10:13:41.696447 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:41 crc kubenswrapper[4966]: I1217 10:13:41.753625 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:41 crc kubenswrapper[4966]: I1217 10:13:41.945473 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:41 crc kubenswrapper[4966]: I1217 10:13:41.945713 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:13:42 crc kubenswrapper[4966]: I1217 10:13:42.992861 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hj8hv" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="registry-server" probeResult="failure" output=< Dec 17 10:13:42 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:13:42 crc kubenswrapper[4966]: > Dec 17 10:13:43 crc kubenswrapper[4966]: I1217 10:13:43.831004 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:13:43 crc kubenswrapper[4966]: E1217 10:13:43.831369 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:13:51 crc kubenswrapper[4966]: I1217 10:13:51.765496 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:51 crc kubenswrapper[4966]: I1217 10:13:51.823631 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvd85"] Dec 17 10:13:51 crc kubenswrapper[4966]: I1217 10:13:51.891671 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qvd85" podUID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerName="registry-server" containerID="cri-o://50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26" gracePeriod=2 Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.519918 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.627062 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-utilities\") pod \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.627184 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zfhl\" (UniqueName: \"kubernetes.io/projected/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-kube-api-access-6zfhl\") pod \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.627219 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-catalog-content\") pod \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\" (UID: \"3e4d1c1d-8170-4d42-b9d3-7128e64766c9\") " Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.631063 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-utilities" (OuterVolumeSpecName: "utilities") pod "3e4d1c1d-8170-4d42-b9d3-7128e64766c9" (UID: "3e4d1c1d-8170-4d42-b9d3-7128e64766c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.642173 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-kube-api-access-6zfhl" (OuterVolumeSpecName: "kube-api-access-6zfhl") pod "3e4d1c1d-8170-4d42-b9d3-7128e64766c9" (UID: "3e4d1c1d-8170-4d42-b9d3-7128e64766c9"). InnerVolumeSpecName "kube-api-access-6zfhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.668700 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e4d1c1d-8170-4d42-b9d3-7128e64766c9" (UID: "3e4d1c1d-8170-4d42-b9d3-7128e64766c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.729693 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.729728 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zfhl\" (UniqueName: \"kubernetes.io/projected/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-kube-api-access-6zfhl\") on node \"crc\" DevicePath \"\"" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.729741 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e4d1c1d-8170-4d42-b9d3-7128e64766c9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.901514 4966 generic.go:334] "Generic (PLEG): container finished" podID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerID="50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26" exitCode=0 Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.901559 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qvd85" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.901556 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvd85" event={"ID":"3e4d1c1d-8170-4d42-b9d3-7128e64766c9","Type":"ContainerDied","Data":"50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26"} Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.901614 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qvd85" event={"ID":"3e4d1c1d-8170-4d42-b9d3-7128e64766c9","Type":"ContainerDied","Data":"cb381ab59ebfc98a18eee5765f4585c5357d8dcdb4eef7e27e1aced90208b2a8"} Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.901632 4966 scope.go:117] "RemoveContainer" containerID="50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.922198 4966 scope.go:117] "RemoveContainer" containerID="6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.935507 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvd85"] Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.954733 4966 scope.go:117] "RemoveContainer" containerID="4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.964007 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qvd85"] Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.991817 4966 scope.go:117] "RemoveContainer" containerID="50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26" Dec 17 10:13:52 crc kubenswrapper[4966]: E1217 10:13:52.993861 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26\": container with ID starting with 50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26 not found: ID does not exist" containerID="50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.994082 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26"} err="failed to get container status \"50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26\": rpc error: code = NotFound desc = could not find container \"50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26\": container with ID starting with 50f1ff3231d1691f9b4909ad200d58c6b50f4844e0d7c8a6b226985865eb5f26 not found: ID does not exist" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.994124 4966 scope.go:117] "RemoveContainer" containerID="6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c" Dec 17 10:13:52 crc kubenswrapper[4966]: E1217 10:13:52.995726 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c\": container with ID starting with 6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c not found: ID does not exist" containerID="6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.995785 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c"} err="failed to get container status \"6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c\": rpc error: code = NotFound desc = could not find container \"6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c\": container with ID starting with 6d05513860b01559309e4cc2eaa152009ca77dfc127d6a52d03ded46a32da66c not found: ID does not exist" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.995820 4966 scope.go:117] "RemoveContainer" containerID="4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8" Dec 17 10:13:52 crc kubenswrapper[4966]: E1217 10:13:52.996438 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8\": container with ID starting with 4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8 not found: ID does not exist" containerID="4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8" Dec 17 10:13:52 crc kubenswrapper[4966]: I1217 10:13:52.996470 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8"} err="failed to get container status \"4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8\": rpc error: code = NotFound desc = could not find container \"4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8\": container with ID starting with 4e92cb8e6edf7c894c6be5c1f776e2bd42d005556853451d0c0ff2f99fbcb7d8 not found: ID does not exist" Dec 17 10:13:53 crc kubenswrapper[4966]: I1217 10:13:53.000137 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hj8hv" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="registry-server" probeResult="failure" output=< Dec 17 10:13:53 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:13:53 crc kubenswrapper[4966]: > Dec 17 10:13:54 crc kubenswrapper[4966]: I1217 10:13:54.842546 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" path="/var/lib/kubelet/pods/3e4d1c1d-8170-4d42-b9d3-7128e64766c9/volumes" Dec 17 10:13:55 crc kubenswrapper[4966]: I1217 10:13:55.830904 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:13:55 crc kubenswrapper[4966]: E1217 10:13:55.831174 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:14:02 crc kubenswrapper[4966]: I1217 10:14:01.999587 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:14:02 crc kubenswrapper[4966]: I1217 10:14:02.056715 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:14:02 crc kubenswrapper[4966]: I1217 10:14:02.570641 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hj8hv"] Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:03.999630 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hj8hv" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="registry-server" containerID="cri-o://0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a" gracePeriod=2 Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.468632 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.596825 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-utilities\") pod \"826113b3-9fc8-41c6-948f-4251cc7c430d\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.597104 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-catalog-content\") pod \"826113b3-9fc8-41c6-948f-4251cc7c430d\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.597135 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfhrx\" (UniqueName: \"kubernetes.io/projected/826113b3-9fc8-41c6-948f-4251cc7c430d-kube-api-access-pfhrx\") pod \"826113b3-9fc8-41c6-948f-4251cc7c430d\" (UID: \"826113b3-9fc8-41c6-948f-4251cc7c430d\") " Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.597579 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-utilities" (OuterVolumeSpecName: "utilities") pod "826113b3-9fc8-41c6-948f-4251cc7c430d" (UID: "826113b3-9fc8-41c6-948f-4251cc7c430d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.597827 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.604527 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/826113b3-9fc8-41c6-948f-4251cc7c430d-kube-api-access-pfhrx" (OuterVolumeSpecName: "kube-api-access-pfhrx") pod "826113b3-9fc8-41c6-948f-4251cc7c430d" (UID: "826113b3-9fc8-41c6-948f-4251cc7c430d"). InnerVolumeSpecName "kube-api-access-pfhrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.700345 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfhrx\" (UniqueName: \"kubernetes.io/projected/826113b3-9fc8-41c6-948f-4251cc7c430d-kube-api-access-pfhrx\") on node \"crc\" DevicePath \"\"" Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.726745 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "826113b3-9fc8-41c6-948f-4251cc7c430d" (UID: "826113b3-9fc8-41c6-948f-4251cc7c430d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:14:04 crc kubenswrapper[4966]: I1217 10:14:04.801788 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/826113b3-9fc8-41c6-948f-4251cc7c430d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.013982 4966 generic.go:334] "Generic (PLEG): container finished" podID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerID="0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a" exitCode=0 Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.014040 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8hv" event={"ID":"826113b3-9fc8-41c6-948f-4251cc7c430d","Type":"ContainerDied","Data":"0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a"} Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.014077 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8hv" event={"ID":"826113b3-9fc8-41c6-948f-4251cc7c430d","Type":"ContainerDied","Data":"e4e6e7b7cc5c2e3ff6cebbf4406d5b3a44d848e88475eb28d2545da090cddfd7"} Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.014110 4966 scope.go:117] "RemoveContainer" containerID="0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.014339 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hj8hv" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.043951 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hj8hv"] Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.048603 4966 scope.go:117] "RemoveContainer" containerID="003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.057601 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hj8hv"] Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.074420 4966 scope.go:117] "RemoveContainer" containerID="943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.129913 4966 scope.go:117] "RemoveContainer" containerID="0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a" Dec 17 10:14:05 crc kubenswrapper[4966]: E1217 10:14:05.135083 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a\": container with ID starting with 0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a not found: ID does not exist" containerID="0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.135209 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a"} err="failed to get container status \"0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a\": rpc error: code = NotFound desc = could not find container \"0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a\": container with ID starting with 0f4629c6f38a13ff05ff89f0c2ece4712a5d054f928ef1ff9912161ac7c0247a not found: ID does not exist" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.135286 4966 scope.go:117] "RemoveContainer" containerID="003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334" Dec 17 10:14:05 crc kubenswrapper[4966]: E1217 10:14:05.135675 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334\": container with ID starting with 003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334 not found: ID does not exist" containerID="003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.135737 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334"} err="failed to get container status \"003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334\": rpc error: code = NotFound desc = could not find container \"003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334\": container with ID starting with 003076ae221c1071a67a9fdf8af9ef74704ea9b70824b041d64c0f42920a3334 not found: ID does not exist" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.135770 4966 scope.go:117] "RemoveContainer" containerID="943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb" Dec 17 10:14:05 crc kubenswrapper[4966]: E1217 10:14:05.136195 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb\": container with ID starting with 943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb not found: ID does not exist" containerID="943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb" Dec 17 10:14:05 crc kubenswrapper[4966]: I1217 10:14:05.136226 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb"} err="failed to get container status \"943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb\": rpc error: code = NotFound desc = could not find container \"943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb\": container with ID starting with 943ed9f6e9e34d8b27a8ddb12ec13189d1356a31c8b8735cd759d1640fe193cb not found: ID does not exist" Dec 17 10:14:06 crc kubenswrapper[4966]: I1217 10:14:06.844307 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" path="/var/lib/kubelet/pods/826113b3-9fc8-41c6-948f-4251cc7c430d/volumes" Dec 17 10:14:08 crc kubenswrapper[4966]: I1217 10:14:08.831709 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:14:08 crc kubenswrapper[4966]: E1217 10:14:08.832427 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:14:20 crc kubenswrapper[4966]: I1217 10:14:20.830561 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:14:20 crc kubenswrapper[4966]: E1217 10:14:20.831323 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:14:32 crc kubenswrapper[4966]: I1217 10:14:32.831184 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:14:32 crc kubenswrapper[4966]: E1217 10:14:32.831832 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:14:44 crc kubenswrapper[4966]: I1217 10:14:44.837133 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:14:44 crc kubenswrapper[4966]: E1217 10:14:44.837892 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:14:56 crc kubenswrapper[4966]: I1217 10:14:56.833090 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:14:56 crc kubenswrapper[4966]: E1217 10:14:56.834202 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.210835 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx"] Dec 17 10:15:00 crc kubenswrapper[4966]: E1217 10:15:00.212053 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerName="registry-server" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.212075 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerName="registry-server" Dec 17 10:15:00 crc kubenswrapper[4966]: E1217 10:15:00.212118 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerName="extract-content" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.212126 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerName="extract-content" Dec 17 10:15:00 crc kubenswrapper[4966]: E1217 10:15:00.212157 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="extract-content" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.212164 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="extract-content" Dec 17 10:15:00 crc kubenswrapper[4966]: E1217 10:15:00.212174 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="registry-server" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.212180 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="registry-server" Dec 17 10:15:00 crc kubenswrapper[4966]: E1217 10:15:00.212192 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerName="extract-utilities" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.212199 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerName="extract-utilities" Dec 17 10:15:00 crc kubenswrapper[4966]: E1217 10:15:00.212217 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="extract-utilities" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.212223 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="extract-utilities" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.212458 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="826113b3-9fc8-41c6-948f-4251cc7c430d" containerName="registry-server" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.212490 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e4d1c1d-8170-4d42-b9d3-7128e64766c9" containerName="registry-server" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.213535 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.278770 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpnjl\" (UniqueName: \"kubernetes.io/projected/1199e280-30bf-4229-9b51-33172ca23479-kube-api-access-vpnjl\") pod \"collect-profiles-29432775-9c6xx\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.278833 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1199e280-30bf-4229-9b51-33172ca23479-secret-volume\") pod \"collect-profiles-29432775-9c6xx\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.278991 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1199e280-30bf-4229-9b51-33172ca23479-config-volume\") pod \"collect-profiles-29432775-9c6xx\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.284261 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.284281 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.306599 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx"] Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.381472 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpnjl\" (UniqueName: \"kubernetes.io/projected/1199e280-30bf-4229-9b51-33172ca23479-kube-api-access-vpnjl\") pod \"collect-profiles-29432775-9c6xx\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.381556 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1199e280-30bf-4229-9b51-33172ca23479-secret-volume\") pod \"collect-profiles-29432775-9c6xx\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.381696 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1199e280-30bf-4229-9b51-33172ca23479-config-volume\") pod \"collect-profiles-29432775-9c6xx\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.382858 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1199e280-30bf-4229-9b51-33172ca23479-config-volume\") pod \"collect-profiles-29432775-9c6xx\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.388473 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1199e280-30bf-4229-9b51-33172ca23479-secret-volume\") pod \"collect-profiles-29432775-9c6xx\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.403435 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpnjl\" (UniqueName: \"kubernetes.io/projected/1199e280-30bf-4229-9b51-33172ca23479-kube-api-access-vpnjl\") pod \"collect-profiles-29432775-9c6xx\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:00 crc kubenswrapper[4966]: I1217 10:15:00.534251 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:01 crc kubenswrapper[4966]: I1217 10:15:01.156904 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx"] Dec 17 10:15:01 crc kubenswrapper[4966]: I1217 10:15:01.524002 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" event={"ID":"1199e280-30bf-4229-9b51-33172ca23479","Type":"ContainerStarted","Data":"22ed0010f84ded624d7c25baa8623747c767714da71173e87eba5ef6160368e1"} Dec 17 10:15:01 crc kubenswrapper[4966]: I1217 10:15:01.524229 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" event={"ID":"1199e280-30bf-4229-9b51-33172ca23479","Type":"ContainerStarted","Data":"73947dd31bec157bf1cfa988595956d8d48745cc6da5e97c132d171961c2bea4"} Dec 17 10:15:01 crc kubenswrapper[4966]: I1217 10:15:01.548629 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" podStartSLOduration=1.54861357 podStartE2EDuration="1.54861357s" podCreationTimestamp="2025-12-17 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 10:15:01.537639921 +0000 UTC m=+6837.082709883" watchObservedRunningTime="2025-12-17 10:15:01.54861357 +0000 UTC m=+6837.093683512" Dec 17 10:15:02 crc kubenswrapper[4966]: I1217 10:15:02.533930 4966 generic.go:334] "Generic (PLEG): container finished" podID="1199e280-30bf-4229-9b51-33172ca23479" containerID="22ed0010f84ded624d7c25baa8623747c767714da71173e87eba5ef6160368e1" exitCode=0 Dec 17 10:15:02 crc kubenswrapper[4966]: I1217 10:15:02.534058 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" event={"ID":"1199e280-30bf-4229-9b51-33172ca23479","Type":"ContainerDied","Data":"22ed0010f84ded624d7c25baa8623747c767714da71173e87eba5ef6160368e1"} Dec 17 10:15:03 crc kubenswrapper[4966]: I1217 10:15:03.910779 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:03 crc kubenswrapper[4966]: I1217 10:15:03.967849 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpnjl\" (UniqueName: \"kubernetes.io/projected/1199e280-30bf-4229-9b51-33172ca23479-kube-api-access-vpnjl\") pod \"1199e280-30bf-4229-9b51-33172ca23479\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " Dec 17 10:15:03 crc kubenswrapper[4966]: I1217 10:15:03.968232 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1199e280-30bf-4229-9b51-33172ca23479-config-volume\") pod \"1199e280-30bf-4229-9b51-33172ca23479\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " Dec 17 10:15:03 crc kubenswrapper[4966]: I1217 10:15:03.968416 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1199e280-30bf-4229-9b51-33172ca23479-secret-volume\") pod \"1199e280-30bf-4229-9b51-33172ca23479\" (UID: \"1199e280-30bf-4229-9b51-33172ca23479\") " Dec 17 10:15:03 crc kubenswrapper[4966]: I1217 10:15:03.968971 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1199e280-30bf-4229-9b51-33172ca23479-config-volume" (OuterVolumeSpecName: "config-volume") pod "1199e280-30bf-4229-9b51-33172ca23479" (UID: "1199e280-30bf-4229-9b51-33172ca23479"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 10:15:03 crc kubenswrapper[4966]: I1217 10:15:03.969224 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1199e280-30bf-4229-9b51-33172ca23479-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 10:15:03 crc kubenswrapper[4966]: I1217 10:15:03.978270 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1199e280-30bf-4229-9b51-33172ca23479-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1199e280-30bf-4229-9b51-33172ca23479" (UID: "1199e280-30bf-4229-9b51-33172ca23479"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:15:03 crc kubenswrapper[4966]: I1217 10:15:03.984524 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1199e280-30bf-4229-9b51-33172ca23479-kube-api-access-vpnjl" (OuterVolumeSpecName: "kube-api-access-vpnjl") pod "1199e280-30bf-4229-9b51-33172ca23479" (UID: "1199e280-30bf-4229-9b51-33172ca23479"). InnerVolumeSpecName "kube-api-access-vpnjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:15:04 crc kubenswrapper[4966]: I1217 10:15:04.071482 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1199e280-30bf-4229-9b51-33172ca23479-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 10:15:04 crc kubenswrapper[4966]: I1217 10:15:04.071522 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpnjl\" (UniqueName: \"kubernetes.io/projected/1199e280-30bf-4229-9b51-33172ca23479-kube-api-access-vpnjl\") on node \"crc\" DevicePath \"\"" Dec 17 10:15:04 crc kubenswrapper[4966]: I1217 10:15:04.550814 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" event={"ID":"1199e280-30bf-4229-9b51-33172ca23479","Type":"ContainerDied","Data":"73947dd31bec157bf1cfa988595956d8d48745cc6da5e97c132d171961c2bea4"} Dec 17 10:15:04 crc kubenswrapper[4966]: I1217 10:15:04.551085 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx" Dec 17 10:15:04 crc kubenswrapper[4966]: I1217 10:15:04.550858 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73947dd31bec157bf1cfa988595956d8d48745cc6da5e97c132d171961c2bea4" Dec 17 10:15:04 crc kubenswrapper[4966]: I1217 10:15:04.638245 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj"] Dec 17 10:15:04 crc kubenswrapper[4966]: I1217 10:15:04.647991 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432730-8jdsj"] Dec 17 10:15:04 crc kubenswrapper[4966]: I1217 10:15:04.841837 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09380640-3b0e-4f60-bfa0-692b6d785749" path="/var/lib/kubelet/pods/09380640-3b0e-4f60-bfa0-692b6d785749/volumes" Dec 17 10:15:09 crc kubenswrapper[4966]: I1217 10:15:09.830465 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:15:09 crc kubenswrapper[4966]: E1217 10:15:09.831041 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:15:23 crc kubenswrapper[4966]: I1217 10:15:23.830321 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:15:23 crc kubenswrapper[4966]: E1217 10:15:23.831794 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:15:35 crc kubenswrapper[4966]: I1217 10:15:35.830606 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:15:35 crc kubenswrapper[4966]: E1217 10:15:35.831638 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:15:49 crc kubenswrapper[4966]: I1217 10:15:49.830600 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:15:49 crc kubenswrapper[4966]: E1217 10:15:49.831396 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:16:00 crc kubenswrapper[4966]: I1217 10:16:00.753753 4966 scope.go:117] "RemoveContainer" containerID="fbb5dbbd73d6d9a85f7e16532d430b24206f1a9b18bde959083b686eeb0d2b22" Dec 17 10:16:03 crc kubenswrapper[4966]: I1217 10:16:03.831175 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:16:03 crc kubenswrapper[4966]: E1217 10:16:03.832924 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:16:14 crc kubenswrapper[4966]: I1217 10:16:14.838852 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:16:14 crc kubenswrapper[4966]: E1217 10:16:14.839679 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:16:27 crc kubenswrapper[4966]: I1217 10:16:27.830476 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:16:27 crc kubenswrapper[4966]: E1217 10:16:27.831201 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:16:38 crc kubenswrapper[4966]: I1217 10:16:38.831737 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:16:38 crc kubenswrapper[4966]: E1217 10:16:38.836286 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.502217 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rgch4"] Dec 17 10:16:44 crc kubenswrapper[4966]: E1217 10:16:44.503263 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1199e280-30bf-4229-9b51-33172ca23479" containerName="collect-profiles" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.503286 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="1199e280-30bf-4229-9b51-33172ca23479" containerName="collect-profiles" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.503649 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="1199e280-30bf-4229-9b51-33172ca23479" containerName="collect-profiles" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.506170 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.524824 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgch4"] Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.699095 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-utilities\") pod \"certified-operators-rgch4\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.699469 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-catalog-content\") pod \"certified-operators-rgch4\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.699509 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mzzh\" (UniqueName: \"kubernetes.io/projected/0c9f239b-d164-4b68-975a-b70de23e3c1a-kube-api-access-4mzzh\") pod \"certified-operators-rgch4\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.800938 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-utilities\") pod \"certified-operators-rgch4\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.801008 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-catalog-content\") pod \"certified-operators-rgch4\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.801042 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mzzh\" (UniqueName: \"kubernetes.io/projected/0c9f239b-d164-4b68-975a-b70de23e3c1a-kube-api-access-4mzzh\") pod \"certified-operators-rgch4\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.801372 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-utilities\") pod \"certified-operators-rgch4\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.801606 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-catalog-content\") pod \"certified-operators-rgch4\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:44 crc kubenswrapper[4966]: I1217 10:16:44.828389 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mzzh\" (UniqueName: \"kubernetes.io/projected/0c9f239b-d164-4b68-975a-b70de23e3c1a-kube-api-access-4mzzh\") pod \"certified-operators-rgch4\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:45 crc kubenswrapper[4966]: I1217 10:16:45.124745 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:45 crc kubenswrapper[4966]: I1217 10:16:45.740346 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgch4"] Dec 17 10:16:46 crc kubenswrapper[4966]: I1217 10:16:46.478925 4966 generic.go:334] "Generic (PLEG): container finished" podID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerID="ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33" exitCode=0 Dec 17 10:16:46 crc kubenswrapper[4966]: I1217 10:16:46.478993 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgch4" event={"ID":"0c9f239b-d164-4b68-975a-b70de23e3c1a","Type":"ContainerDied","Data":"ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33"} Dec 17 10:16:46 crc kubenswrapper[4966]: I1217 10:16:46.479601 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgch4" event={"ID":"0c9f239b-d164-4b68-975a-b70de23e3c1a","Type":"ContainerStarted","Data":"76d80932750d38c4bdf8e6c2c49abb97a5b3f2d7a7809037aaf665669a8176f3"} Dec 17 10:16:48 crc kubenswrapper[4966]: I1217 10:16:48.499584 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgch4" event={"ID":"0c9f239b-d164-4b68-975a-b70de23e3c1a","Type":"ContainerStarted","Data":"92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41"} Dec 17 10:16:49 crc kubenswrapper[4966]: I1217 10:16:49.510245 4966 generic.go:334] "Generic (PLEG): container finished" podID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerID="92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41" exitCode=0 Dec 17 10:16:49 crc kubenswrapper[4966]: I1217 10:16:49.510341 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgch4" event={"ID":"0c9f239b-d164-4b68-975a-b70de23e3c1a","Type":"ContainerDied","Data":"92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41"} Dec 17 10:16:49 crc kubenswrapper[4966]: I1217 10:16:49.830179 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:16:49 crc kubenswrapper[4966]: E1217 10:16:49.830421 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:16:50 crc kubenswrapper[4966]: I1217 10:16:50.526403 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgch4" event={"ID":"0c9f239b-d164-4b68-975a-b70de23e3c1a","Type":"ContainerStarted","Data":"b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c"} Dec 17 10:16:50 crc kubenswrapper[4966]: I1217 10:16:50.552659 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rgch4" podStartSLOduration=2.948178626 podStartE2EDuration="6.552638588s" podCreationTimestamp="2025-12-17 10:16:44 +0000 UTC" firstStartedPulling="2025-12-17 10:16:46.481227819 +0000 UTC m=+6942.026297761" lastFinishedPulling="2025-12-17 10:16:50.085687781 +0000 UTC m=+6945.630757723" observedRunningTime="2025-12-17 10:16:50.550524381 +0000 UTC m=+6946.095594333" watchObservedRunningTime="2025-12-17 10:16:50.552638588 +0000 UTC m=+6946.097708530" Dec 17 10:16:55 crc kubenswrapper[4966]: I1217 10:16:55.125678 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:55 crc kubenswrapper[4966]: I1217 10:16:55.126020 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:16:56 crc kubenswrapper[4966]: I1217 10:16:56.181103 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rgch4" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerName="registry-server" probeResult="failure" output=< Dec 17 10:16:56 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:16:56 crc kubenswrapper[4966]: > Dec 17 10:17:02 crc kubenswrapper[4966]: I1217 10:17:02.830195 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:17:02 crc kubenswrapper[4966]: E1217 10:17:02.831098 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:17:05 crc kubenswrapper[4966]: I1217 10:17:05.176159 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:17:05 crc kubenswrapper[4966]: I1217 10:17:05.232317 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:17:05 crc kubenswrapper[4966]: I1217 10:17:05.440740 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rgch4"] Dec 17 10:17:06 crc kubenswrapper[4966]: I1217 10:17:06.675473 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rgch4" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerName="registry-server" containerID="cri-o://b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c" gracePeriod=2 Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.238568 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.353938 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-catalog-content\") pod \"0c9f239b-d164-4b68-975a-b70de23e3c1a\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.353991 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mzzh\" (UniqueName: \"kubernetes.io/projected/0c9f239b-d164-4b68-975a-b70de23e3c1a-kube-api-access-4mzzh\") pod \"0c9f239b-d164-4b68-975a-b70de23e3c1a\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.354057 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-utilities\") pod \"0c9f239b-d164-4b68-975a-b70de23e3c1a\" (UID: \"0c9f239b-d164-4b68-975a-b70de23e3c1a\") " Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.354578 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-utilities" (OuterVolumeSpecName: "utilities") pod "0c9f239b-d164-4b68-975a-b70de23e3c1a" (UID: "0c9f239b-d164-4b68-975a-b70de23e3c1a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.354888 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.368295 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9f239b-d164-4b68-975a-b70de23e3c1a-kube-api-access-4mzzh" (OuterVolumeSpecName: "kube-api-access-4mzzh") pod "0c9f239b-d164-4b68-975a-b70de23e3c1a" (UID: "0c9f239b-d164-4b68-975a-b70de23e3c1a"). InnerVolumeSpecName "kube-api-access-4mzzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.409799 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c9f239b-d164-4b68-975a-b70de23e3c1a" (UID: "0c9f239b-d164-4b68-975a-b70de23e3c1a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.456258 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c9f239b-d164-4b68-975a-b70de23e3c1a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.456296 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mzzh\" (UniqueName: \"kubernetes.io/projected/0c9f239b-d164-4b68-975a-b70de23e3c1a-kube-api-access-4mzzh\") on node \"crc\" DevicePath \"\"" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.685019 4966 generic.go:334] "Generic (PLEG): container finished" podID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerID="b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c" exitCode=0 Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.685204 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgch4" event={"ID":"0c9f239b-d164-4b68-975a-b70de23e3c1a","Type":"ContainerDied","Data":"b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c"} Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.685350 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgch4" event={"ID":"0c9f239b-d164-4b68-975a-b70de23e3c1a","Type":"ContainerDied","Data":"76d80932750d38c4bdf8e6c2c49abb97a5b3f2d7a7809037aaf665669a8176f3"} Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.685377 4966 scope.go:117] "RemoveContainer" containerID="b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.685285 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgch4" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.727127 4966 scope.go:117] "RemoveContainer" containerID="92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.734892 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rgch4"] Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.758984 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rgch4"] Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.765170 4966 scope.go:117] "RemoveContainer" containerID="ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.806379 4966 scope.go:117] "RemoveContainer" containerID="b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c" Dec 17 10:17:07 crc kubenswrapper[4966]: E1217 10:17:07.807324 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c\": container with ID starting with b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c not found: ID does not exist" containerID="b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.807374 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c"} err="failed to get container status \"b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c\": rpc error: code = NotFound desc = could not find container \"b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c\": container with ID starting with b5e736d2bfdd876e1a01cdb99cfc7768035d2d28d430e8df24c0af3369e6357c not found: ID does not exist" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.807419 4966 scope.go:117] "RemoveContainer" containerID="92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41" Dec 17 10:17:07 crc kubenswrapper[4966]: E1217 10:17:07.811047 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41\": container with ID starting with 92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41 not found: ID does not exist" containerID="92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.811095 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41"} err="failed to get container status \"92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41\": rpc error: code = NotFound desc = could not find container \"92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41\": container with ID starting with 92a3ddf0b0b354cc84d89696c5dc7b272d09972c299b536e07630472f37ccd41 not found: ID does not exist" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.811126 4966 scope.go:117] "RemoveContainer" containerID="ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33" Dec 17 10:17:07 crc kubenswrapper[4966]: E1217 10:17:07.811512 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33\": container with ID starting with ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33 not found: ID does not exist" containerID="ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33" Dec 17 10:17:07 crc kubenswrapper[4966]: I1217 10:17:07.811541 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33"} err="failed to get container status \"ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33\": rpc error: code = NotFound desc = could not find container \"ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33\": container with ID starting with ed81a6766979c3a8e119be5c9b245598a5a5c419ae7404f97a85e377f4097d33 not found: ID does not exist" Dec 17 10:17:08 crc kubenswrapper[4966]: I1217 10:17:08.862046 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" path="/var/lib/kubelet/pods/0c9f239b-d164-4b68-975a-b70de23e3c1a/volumes" Dec 17 10:17:17 crc kubenswrapper[4966]: I1217 10:17:17.831221 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:17:17 crc kubenswrapper[4966]: E1217 10:17:17.832008 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:17:31 crc kubenswrapper[4966]: I1217 10:17:31.830819 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:17:31 crc kubenswrapper[4966]: E1217 10:17:31.831640 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:17:46 crc kubenswrapper[4966]: I1217 10:17:46.832454 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:17:46 crc kubenswrapper[4966]: E1217 10:17:46.833272 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:17:58 crc kubenswrapper[4966]: I1217 10:17:58.830840 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:17:58 crc kubenswrapper[4966]: E1217 10:17:58.831763 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:18:11 crc kubenswrapper[4966]: I1217 10:18:11.831002 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:18:11 crc kubenswrapper[4966]: E1217 10:18:11.831921 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:18:23 crc kubenswrapper[4966]: I1217 10:18:23.830450 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:18:24 crc kubenswrapper[4966]: I1217 10:18:24.388369 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"7639db3cf0fc24de9c4de7b03cf923fff0b666066d0d93da877102d7148f1309"} Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.218113 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j6kc4"] Dec 17 10:18:49 crc kubenswrapper[4966]: E1217 10:18:49.218964 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerName="extract-utilities" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.218977 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerName="extract-utilities" Dec 17 10:18:49 crc kubenswrapper[4966]: E1217 10:18:49.219000 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerName="registry-server" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.219007 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerName="registry-server" Dec 17 10:18:49 crc kubenswrapper[4966]: E1217 10:18:49.219024 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerName="extract-content" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.219031 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerName="extract-content" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.219246 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9f239b-d164-4b68-975a-b70de23e3c1a" containerName="registry-server" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.220790 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.232365 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j6kc4"] Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.367697 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-catalog-content\") pod \"community-operators-j6kc4\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.368266 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-utilities\") pod \"community-operators-j6kc4\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.368439 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9vwr\" (UniqueName: \"kubernetes.io/projected/fd8427e9-fce7-4eb1-9d03-8c63752e9466-kube-api-access-h9vwr\") pod \"community-operators-j6kc4\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.470692 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-catalog-content\") pod \"community-operators-j6kc4\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.470767 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-utilities\") pod \"community-operators-j6kc4\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.470934 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9vwr\" (UniqueName: \"kubernetes.io/projected/fd8427e9-fce7-4eb1-9d03-8c63752e9466-kube-api-access-h9vwr\") pod \"community-operators-j6kc4\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.471218 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-catalog-content\") pod \"community-operators-j6kc4\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.471369 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-utilities\") pod \"community-operators-j6kc4\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.503358 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9vwr\" (UniqueName: \"kubernetes.io/projected/fd8427e9-fce7-4eb1-9d03-8c63752e9466-kube-api-access-h9vwr\") pod \"community-operators-j6kc4\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:49 crc kubenswrapper[4966]: I1217 10:18:49.590647 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:50 crc kubenswrapper[4966]: I1217 10:18:50.131679 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j6kc4"] Dec 17 10:18:50 crc kubenswrapper[4966]: I1217 10:18:50.617640 4966 generic.go:334] "Generic (PLEG): container finished" podID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerID="bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860" exitCode=0 Dec 17 10:18:50 crc kubenswrapper[4966]: I1217 10:18:50.617751 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6kc4" event={"ID":"fd8427e9-fce7-4eb1-9d03-8c63752e9466","Type":"ContainerDied","Data":"bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860"} Dec 17 10:18:50 crc kubenswrapper[4966]: I1217 10:18:50.618023 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6kc4" event={"ID":"fd8427e9-fce7-4eb1-9d03-8c63752e9466","Type":"ContainerStarted","Data":"7461db7d8bb9de6d5d3dfbae05661995b9fe846605d12fce9cd22b4c547ccddd"} Dec 17 10:18:50 crc kubenswrapper[4966]: I1217 10:18:50.620166 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 10:18:52 crc kubenswrapper[4966]: I1217 10:18:52.657310 4966 generic.go:334] "Generic (PLEG): container finished" podID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerID="ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661" exitCode=0 Dec 17 10:18:52 crc kubenswrapper[4966]: I1217 10:18:52.657402 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6kc4" event={"ID":"fd8427e9-fce7-4eb1-9d03-8c63752e9466","Type":"ContainerDied","Data":"ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661"} Dec 17 10:18:53 crc kubenswrapper[4966]: I1217 10:18:53.667465 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6kc4" event={"ID":"fd8427e9-fce7-4eb1-9d03-8c63752e9466","Type":"ContainerStarted","Data":"e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288"} Dec 17 10:18:53 crc kubenswrapper[4966]: I1217 10:18:53.698571 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j6kc4" podStartSLOduration=2.161146802 podStartE2EDuration="4.69855267s" podCreationTimestamp="2025-12-17 10:18:49 +0000 UTC" firstStartedPulling="2025-12-17 10:18:50.619317986 +0000 UTC m=+7066.164387918" lastFinishedPulling="2025-12-17 10:18:53.156723844 +0000 UTC m=+7068.701793786" observedRunningTime="2025-12-17 10:18:53.688345371 +0000 UTC m=+7069.233415323" watchObservedRunningTime="2025-12-17 10:18:53.69855267 +0000 UTC m=+7069.243622612" Dec 17 10:18:59 crc kubenswrapper[4966]: I1217 10:18:59.591704 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:59 crc kubenswrapper[4966]: I1217 10:18:59.593433 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:59 crc kubenswrapper[4966]: I1217 10:18:59.640139 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:59 crc kubenswrapper[4966]: I1217 10:18:59.770289 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:18:59 crc kubenswrapper[4966]: I1217 10:18:59.876799 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j6kc4"] Dec 17 10:19:01 crc kubenswrapper[4966]: I1217 10:19:01.743557 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j6kc4" podUID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerName="registry-server" containerID="cri-o://e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288" gracePeriod=2 Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.264906 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.461649 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-catalog-content\") pod \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.464651 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9vwr\" (UniqueName: \"kubernetes.io/projected/fd8427e9-fce7-4eb1-9d03-8c63752e9466-kube-api-access-h9vwr\") pod \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.464784 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-utilities\") pod \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\" (UID: \"fd8427e9-fce7-4eb1-9d03-8c63752e9466\") " Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.466610 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-utilities" (OuterVolumeSpecName: "utilities") pod "fd8427e9-fce7-4eb1-9d03-8c63752e9466" (UID: "fd8427e9-fce7-4eb1-9d03-8c63752e9466"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.473474 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd8427e9-fce7-4eb1-9d03-8c63752e9466-kube-api-access-h9vwr" (OuterVolumeSpecName: "kube-api-access-h9vwr") pod "fd8427e9-fce7-4eb1-9d03-8c63752e9466" (UID: "fd8427e9-fce7-4eb1-9d03-8c63752e9466"). InnerVolumeSpecName "kube-api-access-h9vwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.519184 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd8427e9-fce7-4eb1-9d03-8c63752e9466" (UID: "fd8427e9-fce7-4eb1-9d03-8c63752e9466"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.567434 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.567468 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9vwr\" (UniqueName: \"kubernetes.io/projected/fd8427e9-fce7-4eb1-9d03-8c63752e9466-kube-api-access-h9vwr\") on node \"crc\" DevicePath \"\"" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.567478 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8427e9-fce7-4eb1-9d03-8c63752e9466-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.756841 4966 generic.go:334] "Generic (PLEG): container finished" podID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerID="e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288" exitCode=0 Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.756944 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6kc4" event={"ID":"fd8427e9-fce7-4eb1-9d03-8c63752e9466","Type":"ContainerDied","Data":"e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288"} Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.756978 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j6kc4" event={"ID":"fd8427e9-fce7-4eb1-9d03-8c63752e9466","Type":"ContainerDied","Data":"7461db7d8bb9de6d5d3dfbae05661995b9fe846605d12fce9cd22b4c547ccddd"} Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.757001 4966 scope.go:117] "RemoveContainer" containerID="e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.757150 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j6kc4" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.800384 4966 scope.go:117] "RemoveContainer" containerID="ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.804480 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j6kc4"] Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.817492 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j6kc4"] Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.827043 4966 scope.go:117] "RemoveContainer" containerID="bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.845476 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" path="/var/lib/kubelet/pods/fd8427e9-fce7-4eb1-9d03-8c63752e9466/volumes" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.883806 4966 scope.go:117] "RemoveContainer" containerID="e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288" Dec 17 10:19:02 crc kubenswrapper[4966]: E1217 10:19:02.884336 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288\": container with ID starting with e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288 not found: ID does not exist" containerID="e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.884382 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288"} err="failed to get container status \"e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288\": rpc error: code = NotFound desc = could not find container \"e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288\": container with ID starting with e492a8296af2964e07e6cb2d73f865ab951512c4201bf3251aae42f124141288 not found: ID does not exist" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.884422 4966 scope.go:117] "RemoveContainer" containerID="ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661" Dec 17 10:19:02 crc kubenswrapper[4966]: E1217 10:19:02.884727 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661\": container with ID starting with ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661 not found: ID does not exist" containerID="ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.884766 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661"} err="failed to get container status \"ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661\": rpc error: code = NotFound desc = could not find container \"ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661\": container with ID starting with ac89d8abd2e52be7bdff08610d0214faaaf8e363aac559b65161e68fbdfd1661 not found: ID does not exist" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.884788 4966 scope.go:117] "RemoveContainer" containerID="bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860" Dec 17 10:19:02 crc kubenswrapper[4966]: E1217 10:19:02.885024 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860\": container with ID starting with bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860 not found: ID does not exist" containerID="bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860" Dec 17 10:19:02 crc kubenswrapper[4966]: I1217 10:19:02.885052 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860"} err="failed to get container status \"bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860\": rpc error: code = NotFound desc = could not find container \"bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860\": container with ID starting with bb260a65d4610b948c8a082305264a8168f1e79f8254a33d403ca51471655860 not found: ID does not exist" Dec 17 10:20:17 crc kubenswrapper[4966]: E1217 10:20:17.554241 4966 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.146:50542->38.102.83.146:36203: write tcp 38.102.83.146:50542->38.102.83.146:36203: write: broken pipe Dec 17 10:20:46 crc kubenswrapper[4966]: I1217 10:20:46.808104 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:20:46 crc kubenswrapper[4966]: I1217 10:20:46.808688 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:21:16 crc kubenswrapper[4966]: I1217 10:21:16.808214 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:21:16 crc kubenswrapper[4966]: I1217 10:21:16.808787 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:21:46 crc kubenswrapper[4966]: I1217 10:21:46.807509 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:21:46 crc kubenswrapper[4966]: I1217 10:21:46.808289 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:21:46 crc kubenswrapper[4966]: I1217 10:21:46.808413 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:21:46 crc kubenswrapper[4966]: I1217 10:21:46.809149 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7639db3cf0fc24de9c4de7b03cf923fff0b666066d0d93da877102d7148f1309"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:21:46 crc kubenswrapper[4966]: I1217 10:21:46.809205 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://7639db3cf0fc24de9c4de7b03cf923fff0b666066d0d93da877102d7148f1309" gracePeriod=600 Dec 17 10:21:47 crc kubenswrapper[4966]: I1217 10:21:47.523818 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="7639db3cf0fc24de9c4de7b03cf923fff0b666066d0d93da877102d7148f1309" exitCode=0 Dec 17 10:21:47 crc kubenswrapper[4966]: I1217 10:21:47.524024 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"7639db3cf0fc24de9c4de7b03cf923fff0b666066d0d93da877102d7148f1309"} Dec 17 10:21:47 crc kubenswrapper[4966]: I1217 10:21:47.524458 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57"} Dec 17 10:21:47 crc kubenswrapper[4966]: I1217 10:21:47.524493 4966 scope.go:117] "RemoveContainer" containerID="c008e028cdd3e25bddae71f45812afbc442addcbcda22b153f8e7208429959d2" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.212414 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vx7d8"] Dec 17 10:23:51 crc kubenswrapper[4966]: E1217 10:23:51.213743 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerName="extract-utilities" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.213765 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerName="extract-utilities" Dec 17 10:23:51 crc kubenswrapper[4966]: E1217 10:23:51.213822 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerName="registry-server" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.213831 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerName="registry-server" Dec 17 10:23:51 crc kubenswrapper[4966]: E1217 10:23:51.213849 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerName="extract-content" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.213858 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerName="extract-content" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.214200 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8427e9-fce7-4eb1-9d03-8c63752e9466" containerName="registry-server" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.241055 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.293255 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx7d8"] Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.394970 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl2ql\" (UniqueName: \"kubernetes.io/projected/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-kube-api-access-dl2ql\") pod \"redhat-marketplace-vx7d8\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.395148 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-utilities\") pod \"redhat-marketplace-vx7d8\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.395205 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-catalog-content\") pod \"redhat-marketplace-vx7d8\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.496488 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-utilities\") pod \"redhat-marketplace-vx7d8\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.496563 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-catalog-content\") pod \"redhat-marketplace-vx7d8\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.496632 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl2ql\" (UniqueName: \"kubernetes.io/projected/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-kube-api-access-dl2ql\") pod \"redhat-marketplace-vx7d8\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.497291 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-utilities\") pod \"redhat-marketplace-vx7d8\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.497453 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-catalog-content\") pod \"redhat-marketplace-vx7d8\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.516197 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl2ql\" (UniqueName: \"kubernetes.io/projected/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-kube-api-access-dl2ql\") pod \"redhat-marketplace-vx7d8\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:51 crc kubenswrapper[4966]: I1217 10:23:51.592019 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:23:52 crc kubenswrapper[4966]: I1217 10:23:52.308085 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx7d8"] Dec 17 10:23:52 crc kubenswrapper[4966]: I1217 10:23:52.742915 4966 generic.go:334] "Generic (PLEG): container finished" podID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerID="45e588b3eefe9a2424a56e7abd4c2eaa671b54c083c86e0d0b47745020f74520" exitCode=0 Dec 17 10:23:52 crc kubenswrapper[4966]: I1217 10:23:52.742967 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx7d8" event={"ID":"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d","Type":"ContainerDied","Data":"45e588b3eefe9a2424a56e7abd4c2eaa671b54c083c86e0d0b47745020f74520"} Dec 17 10:23:52 crc kubenswrapper[4966]: I1217 10:23:52.742999 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx7d8" event={"ID":"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d","Type":"ContainerStarted","Data":"eb9f52460be690c5c1e6969abd1f857010942010778beaa6349205d2eb738262"} Dec 17 10:23:52 crc kubenswrapper[4966]: I1217 10:23:52.745524 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 10:23:53 crc kubenswrapper[4966]: I1217 10:23:53.763190 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx7d8" event={"ID":"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d","Type":"ContainerStarted","Data":"b5dc38fcc6ddff37cd8b6c899b6d6f89da11c6871afc773569b6e9bf1391e173"} Dec 17 10:23:54 crc kubenswrapper[4966]: I1217 10:23:54.775330 4966 generic.go:334] "Generic (PLEG): container finished" podID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerID="b5dc38fcc6ddff37cd8b6c899b6d6f89da11c6871afc773569b6e9bf1391e173" exitCode=0 Dec 17 10:23:54 crc kubenswrapper[4966]: I1217 10:23:54.775655 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx7d8" event={"ID":"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d","Type":"ContainerDied","Data":"b5dc38fcc6ddff37cd8b6c899b6d6f89da11c6871afc773569b6e9bf1391e173"} Dec 17 10:23:55 crc kubenswrapper[4966]: I1217 10:23:55.816932 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx7d8" event={"ID":"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d","Type":"ContainerStarted","Data":"1cc281099aaae8b3b132ca92c2f0bdd0bf1d4f9e7c9a33b81bcd579df7acecc9"} Dec 17 10:23:55 crc kubenswrapper[4966]: I1217 10:23:55.846037 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vx7d8" podStartSLOduration=2.162650398 podStartE2EDuration="4.846018354s" podCreationTimestamp="2025-12-17 10:23:51 +0000 UTC" firstStartedPulling="2025-12-17 10:23:52.745310235 +0000 UTC m=+7368.290380177" lastFinishedPulling="2025-12-17 10:23:55.428678191 +0000 UTC m=+7370.973748133" observedRunningTime="2025-12-17 10:23:55.841261985 +0000 UTC m=+7371.386331937" watchObservedRunningTime="2025-12-17 10:23:55.846018354 +0000 UTC m=+7371.391088296" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.597939 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7stwm"] Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.601447 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.611562 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7stwm"] Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.761914 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-utilities\") pod \"redhat-operators-7stwm\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.761955 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-catalog-content\") pod \"redhat-operators-7stwm\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.762082 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8tfh\" (UniqueName: \"kubernetes.io/projected/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-kube-api-access-k8tfh\") pod \"redhat-operators-7stwm\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.864416 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8tfh\" (UniqueName: \"kubernetes.io/projected/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-kube-api-access-k8tfh\") pod \"redhat-operators-7stwm\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.864817 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-utilities\") pod \"redhat-operators-7stwm\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.864838 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-catalog-content\") pod \"redhat-operators-7stwm\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.865801 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-catalog-content\") pod \"redhat-operators-7stwm\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.866344 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-utilities\") pod \"redhat-operators-7stwm\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.895054 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8tfh\" (UniqueName: \"kubernetes.io/projected/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-kube-api-access-k8tfh\") pod \"redhat-operators-7stwm\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:58 crc kubenswrapper[4966]: I1217 10:23:58.968283 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:23:59 crc kubenswrapper[4966]: I1217 10:23:59.652957 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7stwm"] Dec 17 10:23:59 crc kubenswrapper[4966]: I1217 10:23:59.863142 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7stwm" event={"ID":"4adbbf6e-5276-44d2-8b3e-fb5feec8005a","Type":"ContainerStarted","Data":"d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159"} Dec 17 10:23:59 crc kubenswrapper[4966]: I1217 10:23:59.863478 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7stwm" event={"ID":"4adbbf6e-5276-44d2-8b3e-fb5feec8005a","Type":"ContainerStarted","Data":"510a971d0cf1265eadf78ede661f53068943a04bfb6b2a39d86f9afd226f06c2"} Dec 17 10:24:00 crc kubenswrapper[4966]: I1217 10:24:00.880519 4966 generic.go:334] "Generic (PLEG): container finished" podID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerID="d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159" exitCode=0 Dec 17 10:24:00 crc kubenswrapper[4966]: I1217 10:24:00.880565 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7stwm" event={"ID":"4adbbf6e-5276-44d2-8b3e-fb5feec8005a","Type":"ContainerDied","Data":"d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159"} Dec 17 10:24:01 crc kubenswrapper[4966]: I1217 10:24:01.592183 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:24:01 crc kubenswrapper[4966]: I1217 10:24:01.592654 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:24:01 crc kubenswrapper[4966]: I1217 10:24:01.654136 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:24:01 crc kubenswrapper[4966]: I1217 10:24:01.951809 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:24:02 crc kubenswrapper[4966]: I1217 10:24:02.903316 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7stwm" event={"ID":"4adbbf6e-5276-44d2-8b3e-fb5feec8005a","Type":"ContainerStarted","Data":"3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0"} Dec 17 10:24:04 crc kubenswrapper[4966]: I1217 10:24:04.133733 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx7d8"] Dec 17 10:24:04 crc kubenswrapper[4966]: I1217 10:24:04.134441 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vx7d8" podUID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerName="registry-server" containerID="cri-o://1cc281099aaae8b3b132ca92c2f0bdd0bf1d4f9e7c9a33b81bcd579df7acecc9" gracePeriod=2 Dec 17 10:24:04 crc kubenswrapper[4966]: I1217 10:24:04.924072 4966 generic.go:334] "Generic (PLEG): container finished" podID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerID="1cc281099aaae8b3b132ca92c2f0bdd0bf1d4f9e7c9a33b81bcd579df7acecc9" exitCode=0 Dec 17 10:24:04 crc kubenswrapper[4966]: I1217 10:24:04.924352 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx7d8" event={"ID":"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d","Type":"ContainerDied","Data":"1cc281099aaae8b3b132ca92c2f0bdd0bf1d4f9e7c9a33b81bcd579df7acecc9"} Dec 17 10:24:04 crc kubenswrapper[4966]: I1217 10:24:04.924378 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vx7d8" event={"ID":"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d","Type":"ContainerDied","Data":"eb9f52460be690c5c1e6969abd1f857010942010778beaa6349205d2eb738262"} Dec 17 10:24:04 crc kubenswrapper[4966]: I1217 10:24:04.924388 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb9f52460be690c5c1e6969abd1f857010942010778beaa6349205d2eb738262" Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.030184 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.143285 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-utilities\") pod \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.143377 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-catalog-content\") pod \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.143470 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl2ql\" (UniqueName: \"kubernetes.io/projected/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-kube-api-access-dl2ql\") pod \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\" (UID: \"0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d\") " Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.144190 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-utilities" (OuterVolumeSpecName: "utilities") pod "0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" (UID: "0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.152107 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-kube-api-access-dl2ql" (OuterVolumeSpecName: "kube-api-access-dl2ql") pod "0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" (UID: "0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d"). InnerVolumeSpecName "kube-api-access-dl2ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.162558 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" (UID: "0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.246232 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.246274 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl2ql\" (UniqueName: \"kubernetes.io/projected/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-kube-api-access-dl2ql\") on node \"crc\" DevicePath \"\"" Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.246288 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.934491 4966 generic.go:334] "Generic (PLEG): container finished" podID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerID="3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0" exitCode=0 Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.934851 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vx7d8" Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.934555 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7stwm" event={"ID":"4adbbf6e-5276-44d2-8b3e-fb5feec8005a","Type":"ContainerDied","Data":"3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0"} Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.983989 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx7d8"] Dec 17 10:24:05 crc kubenswrapper[4966]: I1217 10:24:05.994534 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vx7d8"] Dec 17 10:24:06 crc kubenswrapper[4966]: I1217 10:24:06.843211 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" path="/var/lib/kubelet/pods/0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d/volumes" Dec 17 10:24:06 crc kubenswrapper[4966]: I1217 10:24:06.945926 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7stwm" event={"ID":"4adbbf6e-5276-44d2-8b3e-fb5feec8005a","Type":"ContainerStarted","Data":"9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce"} Dec 17 10:24:06 crc kubenswrapper[4966]: I1217 10:24:06.968463 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7stwm" podStartSLOduration=3.3131604599999998 podStartE2EDuration="8.968443404s" podCreationTimestamp="2025-12-17 10:23:58 +0000 UTC" firstStartedPulling="2025-12-17 10:24:00.882831753 +0000 UTC m=+7376.427901695" lastFinishedPulling="2025-12-17 10:24:06.538114697 +0000 UTC m=+7382.083184639" observedRunningTime="2025-12-17 10:24:06.961852484 +0000 UTC m=+7382.506922446" watchObservedRunningTime="2025-12-17 10:24:06.968443404 +0000 UTC m=+7382.513513346" Dec 17 10:24:08 crc kubenswrapper[4966]: I1217 10:24:08.968534 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:24:08 crc kubenswrapper[4966]: I1217 10:24:08.969087 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:24:10 crc kubenswrapper[4966]: I1217 10:24:10.019482 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7stwm" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="registry-server" probeResult="failure" output=< Dec 17 10:24:10 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:24:10 crc kubenswrapper[4966]: > Dec 17 10:24:16 crc kubenswrapper[4966]: I1217 10:24:16.807286 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:24:16 crc kubenswrapper[4966]: I1217 10:24:16.807859 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:24:20 crc kubenswrapper[4966]: I1217 10:24:20.025865 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7stwm" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="registry-server" probeResult="failure" output=< Dec 17 10:24:20 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:24:20 crc kubenswrapper[4966]: > Dec 17 10:24:29 crc kubenswrapper[4966]: I1217 10:24:29.030732 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:24:29 crc kubenswrapper[4966]: I1217 10:24:29.082827 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:24:29 crc kubenswrapper[4966]: I1217 10:24:29.793138 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7stwm"] Dec 17 10:24:30 crc kubenswrapper[4966]: I1217 10:24:30.482084 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7stwm" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="registry-server" containerID="cri-o://9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce" gracePeriod=2 Dec 17 10:24:30 crc kubenswrapper[4966]: I1217 10:24:30.985186 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.147595 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8tfh\" (UniqueName: \"kubernetes.io/projected/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-kube-api-access-k8tfh\") pod \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.147726 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-catalog-content\") pod \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.147895 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-utilities\") pod \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\" (UID: \"4adbbf6e-5276-44d2-8b3e-fb5feec8005a\") " Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.148668 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-utilities" (OuterVolumeSpecName: "utilities") pod "4adbbf6e-5276-44d2-8b3e-fb5feec8005a" (UID: "4adbbf6e-5276-44d2-8b3e-fb5feec8005a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.154476 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-kube-api-access-k8tfh" (OuterVolumeSpecName: "kube-api-access-k8tfh") pod "4adbbf6e-5276-44d2-8b3e-fb5feec8005a" (UID: "4adbbf6e-5276-44d2-8b3e-fb5feec8005a"). InnerVolumeSpecName "kube-api-access-k8tfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.250548 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.250588 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8tfh\" (UniqueName: \"kubernetes.io/projected/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-kube-api-access-k8tfh\") on node \"crc\" DevicePath \"\"" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.257865 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4adbbf6e-5276-44d2-8b3e-fb5feec8005a" (UID: "4adbbf6e-5276-44d2-8b3e-fb5feec8005a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.353635 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4adbbf6e-5276-44d2-8b3e-fb5feec8005a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.497454 4966 generic.go:334] "Generic (PLEG): container finished" podID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerID="9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce" exitCode=0 Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.497503 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7stwm" event={"ID":"4adbbf6e-5276-44d2-8b3e-fb5feec8005a","Type":"ContainerDied","Data":"9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce"} Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.497532 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7stwm" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.497560 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7stwm" event={"ID":"4adbbf6e-5276-44d2-8b3e-fb5feec8005a","Type":"ContainerDied","Data":"510a971d0cf1265eadf78ede661f53068943a04bfb6b2a39d86f9afd226f06c2"} Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.498119 4966 scope.go:117] "RemoveContainer" containerID="9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.521758 4966 scope.go:117] "RemoveContainer" containerID="3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.554073 4966 scope.go:117] "RemoveContainer" containerID="d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.560689 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7stwm"] Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.609973 4966 scope.go:117] "RemoveContainer" containerID="9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce" Dec 17 10:24:31 crc kubenswrapper[4966]: E1217 10:24:31.612311 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce\": container with ID starting with 9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce not found: ID does not exist" containerID="9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.612357 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce"} err="failed to get container status \"9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce\": rpc error: code = NotFound desc = could not find container \"9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce\": container with ID starting with 9821d05a5d7434ca929ce0f7ac184ea3b90a9b0762f21964586d6f877d067fce not found: ID does not exist" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.612385 4966 scope.go:117] "RemoveContainer" containerID="3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0" Dec 17 10:24:31 crc kubenswrapper[4966]: E1217 10:24:31.612727 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0\": container with ID starting with 3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0 not found: ID does not exist" containerID="3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.612744 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0"} err="failed to get container status \"3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0\": rpc error: code = NotFound desc = could not find container \"3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0\": container with ID starting with 3835b60d94d1e97eea80bb71282b432a545153e8081e350ba18d6d0c6886dea0 not found: ID does not exist" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.612764 4966 scope.go:117] "RemoveContainer" containerID="d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159" Dec 17 10:24:31 crc kubenswrapper[4966]: E1217 10:24:31.613062 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159\": container with ID starting with d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159 not found: ID does not exist" containerID="d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.613086 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159"} err="failed to get container status \"d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159\": rpc error: code = NotFound desc = could not find container \"d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159\": container with ID starting with d373b7cb074d7a4093e014e6a00f43b9513a929defe1649bcff0a6b3304b3159 not found: ID does not exist" Dec 17 10:24:31 crc kubenswrapper[4966]: I1217 10:24:31.616145 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7stwm"] Dec 17 10:24:32 crc kubenswrapper[4966]: I1217 10:24:32.845171 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" path="/var/lib/kubelet/pods/4adbbf6e-5276-44d2-8b3e-fb5feec8005a/volumes" Dec 17 10:24:46 crc kubenswrapper[4966]: I1217 10:24:46.807578 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:24:46 crc kubenswrapper[4966]: I1217 10:24:46.808106 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:25:16 crc kubenswrapper[4966]: I1217 10:25:16.808141 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:25:16 crc kubenswrapper[4966]: I1217 10:25:16.808730 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:25:16 crc kubenswrapper[4966]: I1217 10:25:16.808785 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:25:16 crc kubenswrapper[4966]: I1217 10:25:16.809806 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:25:16 crc kubenswrapper[4966]: I1217 10:25:16.809888 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" gracePeriod=600 Dec 17 10:25:16 crc kubenswrapper[4966]: E1217 10:25:16.935414 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:25:17 crc kubenswrapper[4966]: I1217 10:25:17.941364 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" exitCode=0 Dec 17 10:25:17 crc kubenswrapper[4966]: I1217 10:25:17.941418 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57"} Dec 17 10:25:17 crc kubenswrapper[4966]: I1217 10:25:17.941461 4966 scope.go:117] "RemoveContainer" containerID="7639db3cf0fc24de9c4de7b03cf923fff0b666066d0d93da877102d7148f1309" Dec 17 10:25:17 crc kubenswrapper[4966]: I1217 10:25:17.942235 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:25:17 crc kubenswrapper[4966]: E1217 10:25:17.942541 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:25:28 crc kubenswrapper[4966]: I1217 10:25:28.831101 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:25:28 crc kubenswrapper[4966]: E1217 10:25:28.833619 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:25:42 crc kubenswrapper[4966]: I1217 10:25:42.837754 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:25:42 crc kubenswrapper[4966]: E1217 10:25:42.838576 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:25:54 crc kubenswrapper[4966]: I1217 10:25:54.838465 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:25:54 crc kubenswrapper[4966]: E1217 10:25:54.839291 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:26:09 crc kubenswrapper[4966]: I1217 10:26:09.831187 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:26:09 crc kubenswrapper[4966]: E1217 10:26:09.832231 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:26:24 crc kubenswrapper[4966]: I1217 10:26:24.837761 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:26:24 crc kubenswrapper[4966]: E1217 10:26:24.838823 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:26:35 crc kubenswrapper[4966]: I1217 10:26:35.831057 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:26:35 crc kubenswrapper[4966]: E1217 10:26:35.831795 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:26:50 crc kubenswrapper[4966]: I1217 10:26:50.832357 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:26:50 crc kubenswrapper[4966]: E1217 10:26:50.833286 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:27:04 crc kubenswrapper[4966]: I1217 10:27:04.839294 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:27:04 crc kubenswrapper[4966]: E1217 10:27:04.840202 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:27:18 crc kubenswrapper[4966]: I1217 10:27:18.831268 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:27:18 crc kubenswrapper[4966]: E1217 10:27:18.832180 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:27:32 crc kubenswrapper[4966]: I1217 10:27:32.831750 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:27:32 crc kubenswrapper[4966]: E1217 10:27:32.832948 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:27:46 crc kubenswrapper[4966]: I1217 10:27:46.831108 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:27:46 crc kubenswrapper[4966]: E1217 10:27:46.831773 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.116888 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vwzb9"] Dec 17 10:27:53 crc kubenswrapper[4966]: E1217 10:27:53.117962 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerName="extract-utilities" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.117981 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerName="extract-utilities" Dec 17 10:27:53 crc kubenswrapper[4966]: E1217 10:27:53.118003 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerName="registry-server" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.118010 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerName="registry-server" Dec 17 10:27:53 crc kubenswrapper[4966]: E1217 10:27:53.118039 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerName="extract-content" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.118049 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerName="extract-content" Dec 17 10:27:53 crc kubenswrapper[4966]: E1217 10:27:53.118061 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="extract-content" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.118068 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="extract-content" Dec 17 10:27:53 crc kubenswrapper[4966]: E1217 10:27:53.118089 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="registry-server" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.118096 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="registry-server" Dec 17 10:27:53 crc kubenswrapper[4966]: E1217 10:27:53.118109 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="extract-utilities" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.118116 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="extract-utilities" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.118360 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cd6a0a6-1f3d-4a07-8e1b-b2ef6622fa2d" containerName="registry-server" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.118380 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="4adbbf6e-5276-44d2-8b3e-fb5feec8005a" containerName="registry-server" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.120253 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.126644 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vwzb9"] Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.287250 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-catalog-content\") pod \"certified-operators-vwzb9\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.287348 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k58gx\" (UniqueName: \"kubernetes.io/projected/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-kube-api-access-k58gx\") pod \"certified-operators-vwzb9\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.287404 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-utilities\") pod \"certified-operators-vwzb9\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.389048 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k58gx\" (UniqueName: \"kubernetes.io/projected/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-kube-api-access-k58gx\") pod \"certified-operators-vwzb9\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.389162 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-utilities\") pod \"certified-operators-vwzb9\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.389282 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-catalog-content\") pod \"certified-operators-vwzb9\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.389867 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-catalog-content\") pod \"certified-operators-vwzb9\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.389929 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-utilities\") pod \"certified-operators-vwzb9\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.412999 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k58gx\" (UniqueName: \"kubernetes.io/projected/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-kube-api-access-k58gx\") pod \"certified-operators-vwzb9\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.440556 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:27:53 crc kubenswrapper[4966]: I1217 10:27:53.952265 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vwzb9"] Dec 17 10:27:54 crc kubenswrapper[4966]: I1217 10:27:54.615270 4966 generic.go:334] "Generic (PLEG): container finished" podID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerID="e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4" exitCode=0 Dec 17 10:27:54 crc kubenswrapper[4966]: I1217 10:27:54.615323 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwzb9" event={"ID":"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e","Type":"ContainerDied","Data":"e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4"} Dec 17 10:27:54 crc kubenswrapper[4966]: I1217 10:27:54.615781 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwzb9" event={"ID":"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e","Type":"ContainerStarted","Data":"6b7900d07114afa3df5d5cfa76f7a3c062c71cb0a6583a1cfd442b760a50afa3"} Dec 17 10:27:56 crc kubenswrapper[4966]: I1217 10:27:56.646723 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwzb9" event={"ID":"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e","Type":"ContainerStarted","Data":"730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637"} Dec 17 10:27:57 crc kubenswrapper[4966]: I1217 10:27:57.657841 4966 generic.go:334] "Generic (PLEG): container finished" podID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerID="730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637" exitCode=0 Dec 17 10:27:57 crc kubenswrapper[4966]: I1217 10:27:57.657907 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwzb9" event={"ID":"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e","Type":"ContainerDied","Data":"730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637"} Dec 17 10:27:58 crc kubenswrapper[4966]: I1217 10:27:58.830394 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:27:58 crc kubenswrapper[4966]: E1217 10:27:58.834750 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:27:59 crc kubenswrapper[4966]: I1217 10:27:59.678536 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwzb9" event={"ID":"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e","Type":"ContainerStarted","Data":"98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1"} Dec 17 10:27:59 crc kubenswrapper[4966]: I1217 10:27:59.706036 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vwzb9" podStartSLOduration=2.14040767 podStartE2EDuration="6.705796223s" podCreationTimestamp="2025-12-17 10:27:53 +0000 UTC" firstStartedPulling="2025-12-17 10:27:54.617042258 +0000 UTC m=+7610.162112200" lastFinishedPulling="2025-12-17 10:27:59.182430811 +0000 UTC m=+7614.727500753" observedRunningTime="2025-12-17 10:27:59.69981462 +0000 UTC m=+7615.244884572" watchObservedRunningTime="2025-12-17 10:27:59.705796223 +0000 UTC m=+7615.250866165" Dec 17 10:28:03 crc kubenswrapper[4966]: I1217 10:28:03.441661 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:28:03 crc kubenswrapper[4966]: I1217 10:28:03.443200 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:28:04 crc kubenswrapper[4966]: I1217 10:28:04.523365 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-vwzb9" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerName="registry-server" probeResult="failure" output=< Dec 17 10:28:04 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:28:04 crc kubenswrapper[4966]: > Dec 17 10:28:12 crc kubenswrapper[4966]: I1217 10:28:12.832202 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:28:12 crc kubenswrapper[4966]: E1217 10:28:12.833251 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:28:13 crc kubenswrapper[4966]: I1217 10:28:13.501263 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:28:13 crc kubenswrapper[4966]: I1217 10:28:13.557895 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:28:13 crc kubenswrapper[4966]: I1217 10:28:13.751388 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vwzb9"] Dec 17 10:28:14 crc kubenswrapper[4966]: I1217 10:28:14.859972 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vwzb9" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerName="registry-server" containerID="cri-o://98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1" gracePeriod=2 Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.542772 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.694056 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-catalog-content\") pod \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.694129 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-utilities\") pod \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.694217 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k58gx\" (UniqueName: \"kubernetes.io/projected/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-kube-api-access-k58gx\") pod \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\" (UID: \"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e\") " Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.695614 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-utilities" (OuterVolumeSpecName: "utilities") pod "cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" (UID: "cc882d7a-ad73-4f47-a6bb-03b2fabbde1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.721700 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-kube-api-access-k58gx" (OuterVolumeSpecName: "kube-api-access-k58gx") pod "cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" (UID: "cc882d7a-ad73-4f47-a6bb-03b2fabbde1e"). InnerVolumeSpecName "kube-api-access-k58gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.798274 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.798552 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k58gx\" (UniqueName: \"kubernetes.io/projected/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-kube-api-access-k58gx\") on node \"crc\" DevicePath \"\"" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.803251 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" (UID: "cc882d7a-ad73-4f47-a6bb-03b2fabbde1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.868827 4966 generic.go:334] "Generic (PLEG): container finished" podID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerID="98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1" exitCode=0 Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.868908 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwzb9" event={"ID":"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e","Type":"ContainerDied","Data":"98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1"} Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.869752 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwzb9" event={"ID":"cc882d7a-ad73-4f47-a6bb-03b2fabbde1e","Type":"ContainerDied","Data":"6b7900d07114afa3df5d5cfa76f7a3c062c71cb0a6583a1cfd442b760a50afa3"} Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.868959 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vwzb9" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.869821 4966 scope.go:117] "RemoveContainer" containerID="98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.899991 4966 scope.go:117] "RemoveContainer" containerID="730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.900110 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.923753 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vwzb9"] Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.928580 4966 scope.go:117] "RemoveContainer" containerID="e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.942722 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vwzb9"] Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.985626 4966 scope.go:117] "RemoveContainer" containerID="98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1" Dec 17 10:28:15 crc kubenswrapper[4966]: E1217 10:28:15.986253 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1\": container with ID starting with 98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1 not found: ID does not exist" containerID="98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.986287 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1"} err="failed to get container status \"98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1\": rpc error: code = NotFound desc = could not find container \"98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1\": container with ID starting with 98c49ed532c4d0385b8b30be2be0e29a15cf5085bbeff5854fb6a01c086b06d1 not found: ID does not exist" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.986310 4966 scope.go:117] "RemoveContainer" containerID="730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637" Dec 17 10:28:15 crc kubenswrapper[4966]: E1217 10:28:15.987172 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637\": container with ID starting with 730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637 not found: ID does not exist" containerID="730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.987215 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637"} err="failed to get container status \"730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637\": rpc error: code = NotFound desc = could not find container \"730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637\": container with ID starting with 730d5e91de08736c7c381e9e48cf41c57e65810f03b5370fcc09d1fd71507637 not found: ID does not exist" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.987243 4966 scope.go:117] "RemoveContainer" containerID="e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4" Dec 17 10:28:15 crc kubenswrapper[4966]: E1217 10:28:15.987680 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4\": container with ID starting with e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4 not found: ID does not exist" containerID="e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4" Dec 17 10:28:15 crc kubenswrapper[4966]: I1217 10:28:15.988040 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4"} err="failed to get container status \"e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4\": rpc error: code = NotFound desc = could not find container \"e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4\": container with ID starting with e759f8d5ae5d10b70fbb5a2d90d3a5a563e3a235ad99c5efab064a2cf3eb5bc4 not found: ID does not exist" Dec 17 10:28:16 crc kubenswrapper[4966]: I1217 10:28:16.841857 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" path="/var/lib/kubelet/pods/cc882d7a-ad73-4f47-a6bb-03b2fabbde1e/volumes" Dec 17 10:28:23 crc kubenswrapper[4966]: I1217 10:28:23.831416 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:28:23 crc kubenswrapper[4966]: E1217 10:28:23.833213 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:28:34 crc kubenswrapper[4966]: I1217 10:28:34.837056 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:28:34 crc kubenswrapper[4966]: E1217 10:28:34.837906 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:28:47 crc kubenswrapper[4966]: I1217 10:28:47.830735 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:28:47 crc kubenswrapper[4966]: E1217 10:28:47.831657 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:29:01 crc kubenswrapper[4966]: I1217 10:29:01.831033 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:29:01 crc kubenswrapper[4966]: E1217 10:29:01.831801 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:29:13 crc kubenswrapper[4966]: I1217 10:29:13.831320 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:29:13 crc kubenswrapper[4966]: E1217 10:29:13.832324 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:29:27 crc kubenswrapper[4966]: I1217 10:29:27.830930 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:29:27 crc kubenswrapper[4966]: E1217 10:29:27.832021 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.337411 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wvkph"] Dec 17 10:29:33 crc kubenswrapper[4966]: E1217 10:29:33.339644 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerName="extract-content" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.339752 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerName="extract-content" Dec 17 10:29:33 crc kubenswrapper[4966]: E1217 10:29:33.339838 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerName="extract-utilities" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.339937 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerName="extract-utilities" Dec 17 10:29:33 crc kubenswrapper[4966]: E1217 10:29:33.340015 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerName="registry-server" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.340084 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerName="registry-server" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.340431 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc882d7a-ad73-4f47-a6bb-03b2fabbde1e" containerName="registry-server" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.342926 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.371039 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wvkph"] Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.465605 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2986\" (UniqueName: \"kubernetes.io/projected/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-kube-api-access-j2986\") pod \"community-operators-wvkph\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.466296 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-catalog-content\") pod \"community-operators-wvkph\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.466711 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-utilities\") pod \"community-operators-wvkph\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.568523 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-utilities\") pod \"community-operators-wvkph\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.568584 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2986\" (UniqueName: \"kubernetes.io/projected/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-kube-api-access-j2986\") pod \"community-operators-wvkph\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.568647 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-catalog-content\") pod \"community-operators-wvkph\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.568995 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-utilities\") pod \"community-operators-wvkph\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.569132 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-catalog-content\") pod \"community-operators-wvkph\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.587591 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2986\" (UniqueName: \"kubernetes.io/projected/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-kube-api-access-j2986\") pod \"community-operators-wvkph\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:33 crc kubenswrapper[4966]: I1217 10:29:33.663813 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:34 crc kubenswrapper[4966]: I1217 10:29:34.184946 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wvkph"] Dec 17 10:29:34 crc kubenswrapper[4966]: I1217 10:29:34.586079 4966 generic.go:334] "Generic (PLEG): container finished" podID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerID="d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e" exitCode=0 Dec 17 10:29:34 crc kubenswrapper[4966]: I1217 10:29:34.586140 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvkph" event={"ID":"7d16150b-e1f9-4826-a50f-0bc4a2fe886e","Type":"ContainerDied","Data":"d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e"} Dec 17 10:29:34 crc kubenswrapper[4966]: I1217 10:29:34.586459 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvkph" event={"ID":"7d16150b-e1f9-4826-a50f-0bc4a2fe886e","Type":"ContainerStarted","Data":"02c7b52fec411f3b388407927272172e14cc1f95672932549dbdd11e4cfb345d"} Dec 17 10:29:34 crc kubenswrapper[4966]: I1217 10:29:34.588403 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 10:29:36 crc kubenswrapper[4966]: I1217 10:29:36.615595 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvkph" event={"ID":"7d16150b-e1f9-4826-a50f-0bc4a2fe886e","Type":"ContainerStarted","Data":"d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e"} Dec 17 10:29:37 crc kubenswrapper[4966]: I1217 10:29:37.640818 4966 generic.go:334] "Generic (PLEG): container finished" podID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerID="d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e" exitCode=0 Dec 17 10:29:37 crc kubenswrapper[4966]: I1217 10:29:37.641046 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvkph" event={"ID":"7d16150b-e1f9-4826-a50f-0bc4a2fe886e","Type":"ContainerDied","Data":"d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e"} Dec 17 10:29:38 crc kubenswrapper[4966]: I1217 10:29:38.654353 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvkph" event={"ID":"7d16150b-e1f9-4826-a50f-0bc4a2fe886e","Type":"ContainerStarted","Data":"3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44"} Dec 17 10:29:38 crc kubenswrapper[4966]: I1217 10:29:38.680327 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wvkph" podStartSLOduration=1.860959071 podStartE2EDuration="5.680303622s" podCreationTimestamp="2025-12-17 10:29:33 +0000 UTC" firstStartedPulling="2025-12-17 10:29:34.588019093 +0000 UTC m=+7710.133089035" lastFinishedPulling="2025-12-17 10:29:38.407363644 +0000 UTC m=+7713.952433586" observedRunningTime="2025-12-17 10:29:38.673508127 +0000 UTC m=+7714.218578079" watchObservedRunningTime="2025-12-17 10:29:38.680303622 +0000 UTC m=+7714.225373564" Dec 17 10:29:42 crc kubenswrapper[4966]: I1217 10:29:42.831387 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:29:42 crc kubenswrapper[4966]: E1217 10:29:42.831823 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:29:43 crc kubenswrapper[4966]: I1217 10:29:43.664242 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:43 crc kubenswrapper[4966]: I1217 10:29:43.664811 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:43 crc kubenswrapper[4966]: I1217 10:29:43.713643 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:44 crc kubenswrapper[4966]: I1217 10:29:44.757719 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:44 crc kubenswrapper[4966]: I1217 10:29:44.816543 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wvkph"] Dec 17 10:29:46 crc kubenswrapper[4966]: I1217 10:29:46.727290 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wvkph" podUID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerName="registry-server" containerID="cri-o://3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44" gracePeriod=2 Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.283994 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.433828 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-utilities\") pod \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.434103 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2986\" (UniqueName: \"kubernetes.io/projected/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-kube-api-access-j2986\") pod \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.434230 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-catalog-content\") pod \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\" (UID: \"7d16150b-e1f9-4826-a50f-0bc4a2fe886e\") " Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.434828 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-utilities" (OuterVolumeSpecName: "utilities") pod "7d16150b-e1f9-4826-a50f-0bc4a2fe886e" (UID: "7d16150b-e1f9-4826-a50f-0bc4a2fe886e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.434960 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.443229 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-kube-api-access-j2986" (OuterVolumeSpecName: "kube-api-access-j2986") pod "7d16150b-e1f9-4826-a50f-0bc4a2fe886e" (UID: "7d16150b-e1f9-4826-a50f-0bc4a2fe886e"). InnerVolumeSpecName "kube-api-access-j2986". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.508386 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d16150b-e1f9-4826-a50f-0bc4a2fe886e" (UID: "7d16150b-e1f9-4826-a50f-0bc4a2fe886e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.536505 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2986\" (UniqueName: \"kubernetes.io/projected/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-kube-api-access-j2986\") on node \"crc\" DevicePath \"\"" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.536536 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d16150b-e1f9-4826-a50f-0bc4a2fe886e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.740041 4966 generic.go:334] "Generic (PLEG): container finished" podID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerID="3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44" exitCode=0 Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.740088 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvkph" event={"ID":"7d16150b-e1f9-4826-a50f-0bc4a2fe886e","Type":"ContainerDied","Data":"3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44"} Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.740118 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wvkph" event={"ID":"7d16150b-e1f9-4826-a50f-0bc4a2fe886e","Type":"ContainerDied","Data":"02c7b52fec411f3b388407927272172e14cc1f95672932549dbdd11e4cfb345d"} Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.740138 4966 scope.go:117] "RemoveContainer" containerID="3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.740302 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wvkph" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.778144 4966 scope.go:117] "RemoveContainer" containerID="d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.806042 4966 scope.go:117] "RemoveContainer" containerID="d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.807388 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wvkph"] Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.836389 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wvkph"] Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.851967 4966 scope.go:117] "RemoveContainer" containerID="3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44" Dec 17 10:29:47 crc kubenswrapper[4966]: E1217 10:29:47.852724 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44\": container with ID starting with 3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44 not found: ID does not exist" containerID="3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.852792 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44"} err="failed to get container status \"3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44\": rpc error: code = NotFound desc = could not find container \"3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44\": container with ID starting with 3738cb292135c0bf66d55107d9b11643d0f1e216cc617c648aaa6f3527c6cf44 not found: ID does not exist" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.852821 4966 scope.go:117] "RemoveContainer" containerID="d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e" Dec 17 10:29:47 crc kubenswrapper[4966]: E1217 10:29:47.853127 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e\": container with ID starting with d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e not found: ID does not exist" containerID="d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.853157 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e"} err="failed to get container status \"d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e\": rpc error: code = NotFound desc = could not find container \"d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e\": container with ID starting with d035ee057ef2ba74b71323dccf7dd900451c7d83edadf5d2a2d38cac47e7ec4e not found: ID does not exist" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.853178 4966 scope.go:117] "RemoveContainer" containerID="d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e" Dec 17 10:29:47 crc kubenswrapper[4966]: E1217 10:29:47.853483 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e\": container with ID starting with d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e not found: ID does not exist" containerID="d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e" Dec 17 10:29:47 crc kubenswrapper[4966]: I1217 10:29:47.853510 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e"} err="failed to get container status \"d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e\": rpc error: code = NotFound desc = could not find container \"d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e\": container with ID starting with d091769e2c3b7c82ab6e49a02c6a6b4f24073bad2efb74815e4022327f80f46e not found: ID does not exist" Dec 17 10:29:48 crc kubenswrapper[4966]: E1217 10:29:48.012718 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d16150b_e1f9_4826_a50f_0bc4a2fe886e.slice/crio-02c7b52fec411f3b388407927272172e14cc1f95672932549dbdd11e4cfb345d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d16150b_e1f9_4826_a50f_0bc4a2fe886e.slice\": RecentStats: unable to find data in memory cache]" Dec 17 10:29:48 crc kubenswrapper[4966]: I1217 10:29:48.845262 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" path="/var/lib/kubelet/pods/7d16150b-e1f9-4826-a50f-0bc4a2fe886e/volumes" Dec 17 10:29:56 crc kubenswrapper[4966]: I1217 10:29:56.833315 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:29:56 crc kubenswrapper[4966]: E1217 10:29:56.834203 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.263699 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp"] Dec 17 10:30:00 crc kubenswrapper[4966]: E1217 10:30:00.264627 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerName="extract-utilities" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.264652 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerName="extract-utilities" Dec 17 10:30:00 crc kubenswrapper[4966]: E1217 10:30:00.264678 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerName="extract-content" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.264690 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerName="extract-content" Dec 17 10:30:00 crc kubenswrapper[4966]: E1217 10:30:00.264742 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerName="registry-server" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.264755 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerName="registry-server" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.265109 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d16150b-e1f9-4826-a50f-0bc4a2fe886e" containerName="registry-server" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.266684 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.271914 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.271940 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.288616 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp"] Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.379503 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-secret-volume\") pod \"collect-profiles-29432790-44tzp\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.379557 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-config-volume\") pod \"collect-profiles-29432790-44tzp\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.379629 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlkk4\" (UniqueName: \"kubernetes.io/projected/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-kube-api-access-vlkk4\") pod \"collect-profiles-29432790-44tzp\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.482055 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-secret-volume\") pod \"collect-profiles-29432790-44tzp\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.482195 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-config-volume\") pod \"collect-profiles-29432790-44tzp\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.482285 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlkk4\" (UniqueName: \"kubernetes.io/projected/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-kube-api-access-vlkk4\") pod \"collect-profiles-29432790-44tzp\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.483520 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-config-volume\") pod \"collect-profiles-29432790-44tzp\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.488388 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-secret-volume\") pod \"collect-profiles-29432790-44tzp\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.504442 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlkk4\" (UniqueName: \"kubernetes.io/projected/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-kube-api-access-vlkk4\") pod \"collect-profiles-29432790-44tzp\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:00 crc kubenswrapper[4966]: I1217 10:30:00.593018 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:01 crc kubenswrapper[4966]: I1217 10:30:01.049041 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp"] Dec 17 10:30:01 crc kubenswrapper[4966]: I1217 10:30:01.144223 4966 scope.go:117] "RemoveContainer" containerID="b5dc38fcc6ddff37cd8b6c899b6d6f89da11c6871afc773569b6e9bf1391e173" Dec 17 10:30:01 crc kubenswrapper[4966]: I1217 10:30:01.175455 4966 scope.go:117] "RemoveContainer" containerID="45e588b3eefe9a2424a56e7abd4c2eaa671b54c083c86e0d0b47745020f74520" Dec 17 10:30:01 crc kubenswrapper[4966]: I1217 10:30:01.197254 4966 scope.go:117] "RemoveContainer" containerID="1cc281099aaae8b3b132ca92c2f0bdd0bf1d4f9e7c9a33b81bcd579df7acecc9" Dec 17 10:30:01 crc kubenswrapper[4966]: I1217 10:30:01.882012 4966 generic.go:334] "Generic (PLEG): container finished" podID="bb2b6a0e-e006-4016-b0d7-5eca41e881b2" containerID="56649688322a9ae67cfb6512b1c761a4cfb7a786d97f46f2654dad8a155752be" exitCode=0 Dec 17 10:30:01 crc kubenswrapper[4966]: I1217 10:30:01.882057 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" event={"ID":"bb2b6a0e-e006-4016-b0d7-5eca41e881b2","Type":"ContainerDied","Data":"56649688322a9ae67cfb6512b1c761a4cfb7a786d97f46f2654dad8a155752be"} Dec 17 10:30:01 crc kubenswrapper[4966]: I1217 10:30:01.882084 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" event={"ID":"bb2b6a0e-e006-4016-b0d7-5eca41e881b2","Type":"ContainerStarted","Data":"1335cddd6616e2b7e0985140d5b87bae561fe911c3e2ed947aeaf9eec3ed9a22"} Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.246276 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.343508 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlkk4\" (UniqueName: \"kubernetes.io/projected/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-kube-api-access-vlkk4\") pod \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.343935 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-secret-volume\") pod \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.344661 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-config-volume\") pod \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\" (UID: \"bb2b6a0e-e006-4016-b0d7-5eca41e881b2\") " Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.345235 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-config-volume" (OuterVolumeSpecName: "config-volume") pod "bb2b6a0e-e006-4016-b0d7-5eca41e881b2" (UID: "bb2b6a0e-e006-4016-b0d7-5eca41e881b2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.345810 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.349746 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bb2b6a0e-e006-4016-b0d7-5eca41e881b2" (UID: "bb2b6a0e-e006-4016-b0d7-5eca41e881b2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.355211 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-kube-api-access-vlkk4" (OuterVolumeSpecName: "kube-api-access-vlkk4") pod "bb2b6a0e-e006-4016-b0d7-5eca41e881b2" (UID: "bb2b6a0e-e006-4016-b0d7-5eca41e881b2"). InnerVolumeSpecName "kube-api-access-vlkk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.447242 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlkk4\" (UniqueName: \"kubernetes.io/projected/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-kube-api-access-vlkk4\") on node \"crc\" DevicePath \"\"" Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.447280 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb2b6a0e-e006-4016-b0d7-5eca41e881b2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.898496 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" event={"ID":"bb2b6a0e-e006-4016-b0d7-5eca41e881b2","Type":"ContainerDied","Data":"1335cddd6616e2b7e0985140d5b87bae561fe911c3e2ed947aeaf9eec3ed9a22"} Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.898532 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1335cddd6616e2b7e0985140d5b87bae561fe911c3e2ed947aeaf9eec3ed9a22" Dec 17 10:30:03 crc kubenswrapper[4966]: I1217 10:30:03.898585 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp" Dec 17 10:30:04 crc kubenswrapper[4966]: I1217 10:30:04.334317 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6"] Dec 17 10:30:04 crc kubenswrapper[4966]: I1217 10:30:04.344683 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432745-vttc6"] Dec 17 10:30:04 crc kubenswrapper[4966]: I1217 10:30:04.849470 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2670d99-8979-4676-9572-fa1b4112ae40" path="/var/lib/kubelet/pods/f2670d99-8979-4676-9572-fa1b4112ae40/volumes" Dec 17 10:30:07 crc kubenswrapper[4966]: I1217 10:30:07.830679 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:30:07 crc kubenswrapper[4966]: E1217 10:30:07.831142 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:30:22 crc kubenswrapper[4966]: I1217 10:30:22.830973 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:30:23 crc kubenswrapper[4966]: I1217 10:30:23.063250 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"35c011cefd589b659c6fc54977869530e47c7c2344a5001a4bdef4388cfacd63"} Dec 17 10:31:01 crc kubenswrapper[4966]: I1217 10:31:01.264606 4966 scope.go:117] "RemoveContainer" containerID="3079373bd0dbc1ab85c12568dc383adaa32011cb34bd9d67a1fc79cb6108d781" Dec 17 10:32:46 crc kubenswrapper[4966]: I1217 10:32:46.807925 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:32:46 crc kubenswrapper[4966]: I1217 10:32:46.808677 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:33:16 crc kubenswrapper[4966]: I1217 10:33:16.807949 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:33:16 crc kubenswrapper[4966]: I1217 10:33:16.808509 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:33:46 crc kubenswrapper[4966]: I1217 10:33:46.807794 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:33:46 crc kubenswrapper[4966]: I1217 10:33:46.808435 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:33:46 crc kubenswrapper[4966]: I1217 10:33:46.808511 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:33:46 crc kubenswrapper[4966]: I1217 10:33:46.809957 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"35c011cefd589b659c6fc54977869530e47c7c2344a5001a4bdef4388cfacd63"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:33:46 crc kubenswrapper[4966]: I1217 10:33:46.810053 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://35c011cefd589b659c6fc54977869530e47c7c2344a5001a4bdef4388cfacd63" gracePeriod=600 Dec 17 10:33:47 crc kubenswrapper[4966]: I1217 10:33:47.344929 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="35c011cefd589b659c6fc54977869530e47c7c2344a5001a4bdef4388cfacd63" exitCode=0 Dec 17 10:33:47 crc kubenswrapper[4966]: I1217 10:33:47.345180 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"35c011cefd589b659c6fc54977869530e47c7c2344a5001a4bdef4388cfacd63"} Dec 17 10:33:47 crc kubenswrapper[4966]: I1217 10:33:47.345208 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef"} Dec 17 10:33:47 crc kubenswrapper[4966]: I1217 10:33:47.345266 4966 scope.go:117] "RemoveContainer" containerID="3eeaca3313dd47995966f4b3ff22864707dc30e5b6f74b6bcec4655bf1a90b57" Dec 17 10:34:11 crc kubenswrapper[4966]: I1217 10:34:11.836398 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4w2"] Dec 17 10:34:11 crc kubenswrapper[4966]: E1217 10:34:11.837819 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb2b6a0e-e006-4016-b0d7-5eca41e881b2" containerName="collect-profiles" Dec 17 10:34:11 crc kubenswrapper[4966]: I1217 10:34:11.837838 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb2b6a0e-e006-4016-b0d7-5eca41e881b2" containerName="collect-profiles" Dec 17 10:34:11 crc kubenswrapper[4966]: I1217 10:34:11.838380 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb2b6a0e-e006-4016-b0d7-5eca41e881b2" containerName="collect-profiles" Dec 17 10:34:11 crc kubenswrapper[4966]: I1217 10:34:11.847114 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:11 crc kubenswrapper[4966]: I1217 10:34:11.847595 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4w2"] Dec 17 10:34:11 crc kubenswrapper[4966]: I1217 10:34:11.976102 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92994\" (UniqueName: \"kubernetes.io/projected/db243ebf-20ba-4c47-b848-fa6e3eef7c89-kube-api-access-92994\") pod \"redhat-marketplace-bd4w2\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:11 crc kubenswrapper[4966]: I1217 10:34:11.976216 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-catalog-content\") pod \"redhat-marketplace-bd4w2\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:11 crc kubenswrapper[4966]: I1217 10:34:11.976267 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-utilities\") pod \"redhat-marketplace-bd4w2\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:12 crc kubenswrapper[4966]: I1217 10:34:12.078830 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92994\" (UniqueName: \"kubernetes.io/projected/db243ebf-20ba-4c47-b848-fa6e3eef7c89-kube-api-access-92994\") pod \"redhat-marketplace-bd4w2\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:12 crc kubenswrapper[4966]: I1217 10:34:12.079020 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-catalog-content\") pod \"redhat-marketplace-bd4w2\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:12 crc kubenswrapper[4966]: I1217 10:34:12.079106 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-utilities\") pod \"redhat-marketplace-bd4w2\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:12 crc kubenswrapper[4966]: I1217 10:34:12.080149 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-catalog-content\") pod \"redhat-marketplace-bd4w2\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:12 crc kubenswrapper[4966]: I1217 10:34:12.080429 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-utilities\") pod \"redhat-marketplace-bd4w2\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:12 crc kubenswrapper[4966]: I1217 10:34:12.097996 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92994\" (UniqueName: \"kubernetes.io/projected/db243ebf-20ba-4c47-b848-fa6e3eef7c89-kube-api-access-92994\") pod \"redhat-marketplace-bd4w2\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:12 crc kubenswrapper[4966]: I1217 10:34:12.184414 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:12 crc kubenswrapper[4966]: I1217 10:34:12.898930 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4w2"] Dec 17 10:34:13 crc kubenswrapper[4966]: I1217 10:34:13.610371 4966 generic.go:334] "Generic (PLEG): container finished" podID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerID="36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3" exitCode=0 Dec 17 10:34:13 crc kubenswrapper[4966]: I1217 10:34:13.610486 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4w2" event={"ID":"db243ebf-20ba-4c47-b848-fa6e3eef7c89","Type":"ContainerDied","Data":"36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3"} Dec 17 10:34:13 crc kubenswrapper[4966]: I1217 10:34:13.610786 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4w2" event={"ID":"db243ebf-20ba-4c47-b848-fa6e3eef7c89","Type":"ContainerStarted","Data":"60f63a73a071793294dc6acbf77faf4aaab4584cc528319791c179bee627a6a3"} Dec 17 10:34:15 crc kubenswrapper[4966]: I1217 10:34:15.729615 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4w2" event={"ID":"db243ebf-20ba-4c47-b848-fa6e3eef7c89","Type":"ContainerStarted","Data":"9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd"} Dec 17 10:34:16 crc kubenswrapper[4966]: I1217 10:34:16.744894 4966 generic.go:334] "Generic (PLEG): container finished" podID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerID="9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd" exitCode=0 Dec 17 10:34:16 crc kubenswrapper[4966]: I1217 10:34:16.744945 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4w2" event={"ID":"db243ebf-20ba-4c47-b848-fa6e3eef7c89","Type":"ContainerDied","Data":"9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd"} Dec 17 10:34:17 crc kubenswrapper[4966]: I1217 10:34:17.757355 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4w2" event={"ID":"db243ebf-20ba-4c47-b848-fa6e3eef7c89","Type":"ContainerStarted","Data":"7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71"} Dec 17 10:34:17 crc kubenswrapper[4966]: I1217 10:34:17.789107 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bd4w2" podStartSLOduration=3.064111698 podStartE2EDuration="6.788167504s" podCreationTimestamp="2025-12-17 10:34:11 +0000 UTC" firstStartedPulling="2025-12-17 10:34:13.614794664 +0000 UTC m=+7989.159864636" lastFinishedPulling="2025-12-17 10:34:17.3388505 +0000 UTC m=+7992.883920442" observedRunningTime="2025-12-17 10:34:17.77703307 +0000 UTC m=+7993.322103012" watchObservedRunningTime="2025-12-17 10:34:17.788167504 +0000 UTC m=+7993.333237456" Dec 17 10:34:22 crc kubenswrapper[4966]: I1217 10:34:22.185272 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:22 crc kubenswrapper[4966]: I1217 10:34:22.185775 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:22 crc kubenswrapper[4966]: I1217 10:34:22.274764 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:22 crc kubenswrapper[4966]: I1217 10:34:22.878988 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:22 crc kubenswrapper[4966]: I1217 10:34:22.937052 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4w2"] Dec 17 10:34:24 crc kubenswrapper[4966]: I1217 10:34:24.874529 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bd4w2" podUID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerName="registry-server" containerID="cri-o://7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71" gracePeriod=2 Dec 17 10:34:25 crc kubenswrapper[4966]: E1217 10:34:25.121618 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb243ebf_20ba_4c47_b848_fa6e3eef7c89.slice/crio-7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb243ebf_20ba_4c47_b848_fa6e3eef7c89.slice/crio-conmon-7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71.scope\": RecentStats: unable to find data in memory cache]" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.403579 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.493935 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92994\" (UniqueName: \"kubernetes.io/projected/db243ebf-20ba-4c47-b848-fa6e3eef7c89-kube-api-access-92994\") pod \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.494027 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-catalog-content\") pod \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.494168 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-utilities\") pod \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\" (UID: \"db243ebf-20ba-4c47-b848-fa6e3eef7c89\") " Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.495201 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-utilities" (OuterVolumeSpecName: "utilities") pod "db243ebf-20ba-4c47-b848-fa6e3eef7c89" (UID: "db243ebf-20ba-4c47-b848-fa6e3eef7c89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.502100 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db243ebf-20ba-4c47-b848-fa6e3eef7c89-kube-api-access-92994" (OuterVolumeSpecName: "kube-api-access-92994") pod "db243ebf-20ba-4c47-b848-fa6e3eef7c89" (UID: "db243ebf-20ba-4c47-b848-fa6e3eef7c89"). InnerVolumeSpecName "kube-api-access-92994". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.518349 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db243ebf-20ba-4c47-b848-fa6e3eef7c89" (UID: "db243ebf-20ba-4c47-b848-fa6e3eef7c89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.596243 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92994\" (UniqueName: \"kubernetes.io/projected/db243ebf-20ba-4c47-b848-fa6e3eef7c89-kube-api-access-92994\") on node \"crc\" DevicePath \"\"" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.596466 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.596535 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db243ebf-20ba-4c47-b848-fa6e3eef7c89-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.886011 4966 generic.go:334] "Generic (PLEG): container finished" podID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerID="7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71" exitCode=0 Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.886063 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4w2" event={"ID":"db243ebf-20ba-4c47-b848-fa6e3eef7c89","Type":"ContainerDied","Data":"7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71"} Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.886091 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4w2" event={"ID":"db243ebf-20ba-4c47-b848-fa6e3eef7c89","Type":"ContainerDied","Data":"60f63a73a071793294dc6acbf77faf4aaab4584cc528319791c179bee627a6a3"} Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.886112 4966 scope.go:117] "RemoveContainer" containerID="7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.886145 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd4w2" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.909123 4966 scope.go:117] "RemoveContainer" containerID="9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.942741 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4w2"] Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.945213 4966 scope.go:117] "RemoveContainer" containerID="36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.953940 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4w2"] Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.980144 4966 scope.go:117] "RemoveContainer" containerID="7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71" Dec 17 10:34:25 crc kubenswrapper[4966]: E1217 10:34:25.981285 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71\": container with ID starting with 7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71 not found: ID does not exist" containerID="7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.981322 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71"} err="failed to get container status \"7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71\": rpc error: code = NotFound desc = could not find container \"7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71\": container with ID starting with 7bdb27d1312ea6403303722b1203fb0a786a7811161f05f7bb4532d251916d71 not found: ID does not exist" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.981357 4966 scope.go:117] "RemoveContainer" containerID="9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd" Dec 17 10:34:25 crc kubenswrapper[4966]: E1217 10:34:25.981617 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd\": container with ID starting with 9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd not found: ID does not exist" containerID="9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.981652 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd"} err="failed to get container status \"9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd\": rpc error: code = NotFound desc = could not find container \"9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd\": container with ID starting with 9649a9e9f849738401b98e711622da6bc19e5e719fcc8788aadd3b3c8b6a43dd not found: ID does not exist" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.981664 4966 scope.go:117] "RemoveContainer" containerID="36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3" Dec 17 10:34:25 crc kubenswrapper[4966]: E1217 10:34:25.982043 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3\": container with ID starting with 36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3 not found: ID does not exist" containerID="36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3" Dec 17 10:34:25 crc kubenswrapper[4966]: I1217 10:34:25.982067 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3"} err="failed to get container status \"36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3\": rpc error: code = NotFound desc = could not find container \"36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3\": container with ID starting with 36013c7e9c2251aa400d94cd106eb61b97a33172c68e3dce6f5ff55bdd8fbcd3 not found: ID does not exist" Dec 17 10:34:26 crc kubenswrapper[4966]: I1217 10:34:26.842035 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" path="/var/lib/kubelet/pods/db243ebf-20ba-4c47-b848-fa6e3eef7c89/volumes" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.184239 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7cmzh"] Dec 17 10:34:51 crc kubenswrapper[4966]: E1217 10:34:51.185033 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerName="extract-utilities" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.185045 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerName="extract-utilities" Dec 17 10:34:51 crc kubenswrapper[4966]: E1217 10:34:51.185058 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerName="extract-content" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.185063 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerName="extract-content" Dec 17 10:34:51 crc kubenswrapper[4966]: E1217 10:34:51.185073 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerName="registry-server" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.185079 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerName="registry-server" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.185281 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="db243ebf-20ba-4c47-b848-fa6e3eef7c89" containerName="registry-server" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.186608 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.213387 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7cmzh"] Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.306161 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmjw2\" (UniqueName: \"kubernetes.io/projected/4a9bca64-a38e-498f-8343-a01b3c39009a-kube-api-access-dmjw2\") pod \"redhat-operators-7cmzh\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.306256 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-utilities\") pod \"redhat-operators-7cmzh\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.306296 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-catalog-content\") pod \"redhat-operators-7cmzh\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.407461 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmjw2\" (UniqueName: \"kubernetes.io/projected/4a9bca64-a38e-498f-8343-a01b3c39009a-kube-api-access-dmjw2\") pod \"redhat-operators-7cmzh\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.407807 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-utilities\") pod \"redhat-operators-7cmzh\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.407849 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-catalog-content\") pod \"redhat-operators-7cmzh\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.408480 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-utilities\") pod \"redhat-operators-7cmzh\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.408515 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-catalog-content\") pod \"redhat-operators-7cmzh\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.426150 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmjw2\" (UniqueName: \"kubernetes.io/projected/4a9bca64-a38e-498f-8343-a01b3c39009a-kube-api-access-dmjw2\") pod \"redhat-operators-7cmzh\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:51 crc kubenswrapper[4966]: I1217 10:34:51.519745 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:34:52 crc kubenswrapper[4966]: I1217 10:34:52.026509 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7cmzh"] Dec 17 10:34:52 crc kubenswrapper[4966]: I1217 10:34:52.143255 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cmzh" event={"ID":"4a9bca64-a38e-498f-8343-a01b3c39009a","Type":"ContainerStarted","Data":"43b6ae482b493a3dc363ef865b3539bf2625cef0116fb82b36523456eb3d252e"} Dec 17 10:34:53 crc kubenswrapper[4966]: I1217 10:34:53.154371 4966 generic.go:334] "Generic (PLEG): container finished" podID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerID="084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5" exitCode=0 Dec 17 10:34:53 crc kubenswrapper[4966]: I1217 10:34:53.154461 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cmzh" event={"ID":"4a9bca64-a38e-498f-8343-a01b3c39009a","Type":"ContainerDied","Data":"084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5"} Dec 17 10:34:53 crc kubenswrapper[4966]: I1217 10:34:53.156842 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 10:34:54 crc kubenswrapper[4966]: I1217 10:34:54.166699 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cmzh" event={"ID":"4a9bca64-a38e-498f-8343-a01b3c39009a","Type":"ContainerStarted","Data":"cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678"} Dec 17 10:34:57 crc kubenswrapper[4966]: I1217 10:34:57.194269 4966 generic.go:334] "Generic (PLEG): container finished" podID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerID="cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678" exitCode=0 Dec 17 10:34:57 crc kubenswrapper[4966]: I1217 10:34:57.194360 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cmzh" event={"ID":"4a9bca64-a38e-498f-8343-a01b3c39009a","Type":"ContainerDied","Data":"cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678"} Dec 17 10:34:58 crc kubenswrapper[4966]: I1217 10:34:58.205050 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cmzh" event={"ID":"4a9bca64-a38e-498f-8343-a01b3c39009a","Type":"ContainerStarted","Data":"f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e"} Dec 17 10:34:58 crc kubenswrapper[4966]: I1217 10:34:58.227794 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7cmzh" podStartSLOduration=2.497718766 podStartE2EDuration="7.227772596s" podCreationTimestamp="2025-12-17 10:34:51 +0000 UTC" firstStartedPulling="2025-12-17 10:34:53.156459356 +0000 UTC m=+8028.701529298" lastFinishedPulling="2025-12-17 10:34:57.886513166 +0000 UTC m=+8033.431583128" observedRunningTime="2025-12-17 10:34:58.22390974 +0000 UTC m=+8033.768979692" watchObservedRunningTime="2025-12-17 10:34:58.227772596 +0000 UTC m=+8033.772842538" Dec 17 10:35:01 crc kubenswrapper[4966]: I1217 10:35:01.520244 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:35:01 crc kubenswrapper[4966]: I1217 10:35:01.520761 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:35:02 crc kubenswrapper[4966]: I1217 10:35:02.570320 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7cmzh" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="registry-server" probeResult="failure" output=< Dec 17 10:35:02 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:35:02 crc kubenswrapper[4966]: > Dec 17 10:35:12 crc kubenswrapper[4966]: I1217 10:35:12.571665 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7cmzh" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="registry-server" probeResult="failure" output=< Dec 17 10:35:12 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:35:12 crc kubenswrapper[4966]: > Dec 17 10:35:21 crc kubenswrapper[4966]: I1217 10:35:21.593822 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:35:21 crc kubenswrapper[4966]: I1217 10:35:21.649938 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:35:22 crc kubenswrapper[4966]: I1217 10:35:22.192101 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7cmzh"] Dec 17 10:35:23 crc kubenswrapper[4966]: I1217 10:35:23.466548 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7cmzh" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="registry-server" containerID="cri-o://f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e" gracePeriod=2 Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.021946 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.224391 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-utilities\") pod \"4a9bca64-a38e-498f-8343-a01b3c39009a\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.225218 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-utilities" (OuterVolumeSpecName: "utilities") pod "4a9bca64-a38e-498f-8343-a01b3c39009a" (UID: "4a9bca64-a38e-498f-8343-a01b3c39009a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.225499 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-catalog-content\") pod \"4a9bca64-a38e-498f-8343-a01b3c39009a\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.225719 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmjw2\" (UniqueName: \"kubernetes.io/projected/4a9bca64-a38e-498f-8343-a01b3c39009a-kube-api-access-dmjw2\") pod \"4a9bca64-a38e-498f-8343-a01b3c39009a\" (UID: \"4a9bca64-a38e-498f-8343-a01b3c39009a\") " Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.228081 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.232135 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a9bca64-a38e-498f-8343-a01b3c39009a-kube-api-access-dmjw2" (OuterVolumeSpecName: "kube-api-access-dmjw2") pod "4a9bca64-a38e-498f-8343-a01b3c39009a" (UID: "4a9bca64-a38e-498f-8343-a01b3c39009a"). InnerVolumeSpecName "kube-api-access-dmjw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.330631 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmjw2\" (UniqueName: \"kubernetes.io/projected/4a9bca64-a38e-498f-8343-a01b3c39009a-kube-api-access-dmjw2\") on node \"crc\" DevicePath \"\"" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.336006 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a9bca64-a38e-498f-8343-a01b3c39009a" (UID: "4a9bca64-a38e-498f-8343-a01b3c39009a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.432053 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a9bca64-a38e-498f-8343-a01b3c39009a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.475520 4966 generic.go:334] "Generic (PLEG): container finished" podID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerID="f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e" exitCode=0 Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.475581 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cmzh" event={"ID":"4a9bca64-a38e-498f-8343-a01b3c39009a","Type":"ContainerDied","Data":"f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e"} Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.475595 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7cmzh" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.475607 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7cmzh" event={"ID":"4a9bca64-a38e-498f-8343-a01b3c39009a","Type":"ContainerDied","Data":"43b6ae482b493a3dc363ef865b3539bf2625cef0116fb82b36523456eb3d252e"} Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.475647 4966 scope.go:117] "RemoveContainer" containerID="f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.497296 4966 scope.go:117] "RemoveContainer" containerID="cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.521177 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7cmzh"] Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.535453 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7cmzh"] Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.538303 4966 scope.go:117] "RemoveContainer" containerID="084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.576663 4966 scope.go:117] "RemoveContainer" containerID="f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e" Dec 17 10:35:24 crc kubenswrapper[4966]: E1217 10:35:24.577024 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e\": container with ID starting with f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e not found: ID does not exist" containerID="f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.577065 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e"} err="failed to get container status \"f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e\": rpc error: code = NotFound desc = could not find container \"f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e\": container with ID starting with f7a79a4550f4b5066a9adadf6a0b7581880672f9452dfbea30ce687e3826e07e not found: ID does not exist" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.577089 4966 scope.go:117] "RemoveContainer" containerID="cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678" Dec 17 10:35:24 crc kubenswrapper[4966]: E1217 10:35:24.577473 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678\": container with ID starting with cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678 not found: ID does not exist" containerID="cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.577509 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678"} err="failed to get container status \"cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678\": rpc error: code = NotFound desc = could not find container \"cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678\": container with ID starting with cb5d7dfdcbf0e0ed06b9927356a97735ca9417ee9822c90d684522df74b60678 not found: ID does not exist" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.577528 4966 scope.go:117] "RemoveContainer" containerID="084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5" Dec 17 10:35:24 crc kubenswrapper[4966]: E1217 10:35:24.577790 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5\": container with ID starting with 084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5 not found: ID does not exist" containerID="084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.577816 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5"} err="failed to get container status \"084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5\": rpc error: code = NotFound desc = could not find container \"084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5\": container with ID starting with 084194db4b404324f5200db5582a97c5af06b4441cf7731e4b6b073e38f86ae5 not found: ID does not exist" Dec 17 10:35:24 crc kubenswrapper[4966]: I1217 10:35:24.841040 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" path="/var/lib/kubelet/pods/4a9bca64-a38e-498f-8343-a01b3c39009a/volumes" Dec 17 10:36:16 crc kubenswrapper[4966]: I1217 10:36:16.808105 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:36:16 crc kubenswrapper[4966]: I1217 10:36:16.810190 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:36:25 crc kubenswrapper[4966]: E1217 10:36:25.546614 4966 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.146:41528->38.102.83.146:36203: write tcp 38.102.83.146:41528->38.102.83.146:36203: write: broken pipe Dec 17 10:36:46 crc kubenswrapper[4966]: I1217 10:36:46.808423 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:36:46 crc kubenswrapper[4966]: I1217 10:36:46.810494 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:37:16 crc kubenswrapper[4966]: I1217 10:37:16.808300 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:37:16 crc kubenswrapper[4966]: I1217 10:37:16.808858 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:37:16 crc kubenswrapper[4966]: I1217 10:37:16.808937 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:37:16 crc kubenswrapper[4966]: I1217 10:37:16.809716 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:37:16 crc kubenswrapper[4966]: I1217 10:37:16.809786 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" gracePeriod=600 Dec 17 10:37:16 crc kubenswrapper[4966]: E1217 10:37:16.935324 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:37:17 crc kubenswrapper[4966]: I1217 10:37:17.526578 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" exitCode=0 Dec 17 10:37:17 crc kubenswrapper[4966]: I1217 10:37:17.526627 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef"} Dec 17 10:37:17 crc kubenswrapper[4966]: I1217 10:37:17.526700 4966 scope.go:117] "RemoveContainer" containerID="35c011cefd589b659c6fc54977869530e47c7c2344a5001a4bdef4388cfacd63" Dec 17 10:37:17 crc kubenswrapper[4966]: I1217 10:37:17.527505 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:37:17 crc kubenswrapper[4966]: E1217 10:37:17.527784 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:37:30 crc kubenswrapper[4966]: I1217 10:37:30.831145 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:37:30 crc kubenswrapper[4966]: E1217 10:37:30.831748 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:37:43 crc kubenswrapper[4966]: I1217 10:37:43.830510 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:37:43 crc kubenswrapper[4966]: E1217 10:37:43.831441 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:37:57 crc kubenswrapper[4966]: I1217 10:37:57.830427 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:37:57 crc kubenswrapper[4966]: E1217 10:37:57.831206 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:38:12 crc kubenswrapper[4966]: I1217 10:38:12.830620 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:38:12 crc kubenswrapper[4966]: E1217 10:38:12.831502 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:38:26 crc kubenswrapper[4966]: I1217 10:38:26.831446 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:38:26 crc kubenswrapper[4966]: E1217 10:38:26.832415 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:38:39 crc kubenswrapper[4966]: I1217 10:38:39.831077 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:38:39 crc kubenswrapper[4966]: E1217 10:38:39.832325 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:38:51 crc kubenswrapper[4966]: I1217 10:38:51.833660 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:38:51 crc kubenswrapper[4966]: E1217 10:38:51.834668 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:39:02 crc kubenswrapper[4966]: I1217 10:39:02.830450 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:39:02 crc kubenswrapper[4966]: E1217 10:39:02.831327 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:39:13 crc kubenswrapper[4966]: I1217 10:39:13.832793 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:39:13 crc kubenswrapper[4966]: E1217 10:39:13.833760 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.094101 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q92zd"] Dec 17 10:39:15 crc kubenswrapper[4966]: E1217 10:39:15.096264 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="registry-server" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.096368 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="registry-server" Dec 17 10:39:15 crc kubenswrapper[4966]: E1217 10:39:15.096432 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="extract-content" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.096486 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="extract-content" Dec 17 10:39:15 crc kubenswrapper[4966]: E1217 10:39:15.096549 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="extract-utilities" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.096602 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="extract-utilities" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.097204 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a9bca64-a38e-498f-8343-a01b3c39009a" containerName="registry-server" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.099670 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.137940 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q92zd"] Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.225911 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-catalog-content\") pod \"certified-operators-q92zd\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.226079 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-utilities\") pod \"certified-operators-q92zd\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.226137 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w84t\" (UniqueName: \"kubernetes.io/projected/3afec990-7323-4304-997d-cb2ccfd9a39b-kube-api-access-5w84t\") pod \"certified-operators-q92zd\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.329426 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-utilities\") pod \"certified-operators-q92zd\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.329494 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w84t\" (UniqueName: \"kubernetes.io/projected/3afec990-7323-4304-997d-cb2ccfd9a39b-kube-api-access-5w84t\") pod \"certified-operators-q92zd\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.329608 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-catalog-content\") pod \"certified-operators-q92zd\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.330166 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-catalog-content\") pod \"certified-operators-q92zd\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.330403 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-utilities\") pod \"certified-operators-q92zd\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.350422 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w84t\" (UniqueName: \"kubernetes.io/projected/3afec990-7323-4304-997d-cb2ccfd9a39b-kube-api-access-5w84t\") pod \"certified-operators-q92zd\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.426072 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:15 crc kubenswrapper[4966]: I1217 10:39:15.983149 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q92zd"] Dec 17 10:39:16 crc kubenswrapper[4966]: I1217 10:39:16.651213 4966 generic.go:334] "Generic (PLEG): container finished" podID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerID="73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104" exitCode=0 Dec 17 10:39:16 crc kubenswrapper[4966]: I1217 10:39:16.651307 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q92zd" event={"ID":"3afec990-7323-4304-997d-cb2ccfd9a39b","Type":"ContainerDied","Data":"73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104"} Dec 17 10:39:16 crc kubenswrapper[4966]: I1217 10:39:16.651591 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q92zd" event={"ID":"3afec990-7323-4304-997d-cb2ccfd9a39b","Type":"ContainerStarted","Data":"328fa47fc0d8f122bab40a91500cd7922a1cb33e51c20fd8eb9963c0c84b3d90"} Dec 17 10:39:18 crc kubenswrapper[4966]: I1217 10:39:18.692557 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q92zd" event={"ID":"3afec990-7323-4304-997d-cb2ccfd9a39b","Type":"ContainerStarted","Data":"45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6"} Dec 17 10:39:19 crc kubenswrapper[4966]: I1217 10:39:19.704040 4966 generic.go:334] "Generic (PLEG): container finished" podID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerID="45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6" exitCode=0 Dec 17 10:39:19 crc kubenswrapper[4966]: I1217 10:39:19.704252 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q92zd" event={"ID":"3afec990-7323-4304-997d-cb2ccfd9a39b","Type":"ContainerDied","Data":"45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6"} Dec 17 10:39:20 crc kubenswrapper[4966]: I1217 10:39:20.713281 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q92zd" event={"ID":"3afec990-7323-4304-997d-cb2ccfd9a39b","Type":"ContainerStarted","Data":"94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c"} Dec 17 10:39:20 crc kubenswrapper[4966]: I1217 10:39:20.731628 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q92zd" podStartSLOduration=2.072164167 podStartE2EDuration="5.731226772s" podCreationTimestamp="2025-12-17 10:39:15 +0000 UTC" firstStartedPulling="2025-12-17 10:39:16.653290612 +0000 UTC m=+8292.198360554" lastFinishedPulling="2025-12-17 10:39:20.312353227 +0000 UTC m=+8295.857423159" observedRunningTime="2025-12-17 10:39:20.727352676 +0000 UTC m=+8296.272422628" watchObservedRunningTime="2025-12-17 10:39:20.731226772 +0000 UTC m=+8296.276296714" Dec 17 10:39:25 crc kubenswrapper[4966]: I1217 10:39:25.426583 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:25 crc kubenswrapper[4966]: I1217 10:39:25.427226 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:26 crc kubenswrapper[4966]: I1217 10:39:26.471256 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-q92zd" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerName="registry-server" probeResult="failure" output=< Dec 17 10:39:26 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:39:26 crc kubenswrapper[4966]: > Dec 17 10:39:26 crc kubenswrapper[4966]: I1217 10:39:26.830511 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:39:26 crc kubenswrapper[4966]: E1217 10:39:26.830982 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:39:35 crc kubenswrapper[4966]: I1217 10:39:35.479515 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:35 crc kubenswrapper[4966]: I1217 10:39:35.544994 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:35 crc kubenswrapper[4966]: I1217 10:39:35.719751 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q92zd"] Dec 17 10:39:36 crc kubenswrapper[4966]: I1217 10:39:36.875819 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q92zd" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerName="registry-server" containerID="cri-o://94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c" gracePeriod=2 Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.353023 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.393292 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-utilities\") pod \"3afec990-7323-4304-997d-cb2ccfd9a39b\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.393370 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-catalog-content\") pod \"3afec990-7323-4304-997d-cb2ccfd9a39b\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.393473 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w84t\" (UniqueName: \"kubernetes.io/projected/3afec990-7323-4304-997d-cb2ccfd9a39b-kube-api-access-5w84t\") pod \"3afec990-7323-4304-997d-cb2ccfd9a39b\" (UID: \"3afec990-7323-4304-997d-cb2ccfd9a39b\") " Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.394627 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-utilities" (OuterVolumeSpecName: "utilities") pod "3afec990-7323-4304-997d-cb2ccfd9a39b" (UID: "3afec990-7323-4304-997d-cb2ccfd9a39b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.424114 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3afec990-7323-4304-997d-cb2ccfd9a39b-kube-api-access-5w84t" (OuterVolumeSpecName: "kube-api-access-5w84t") pod "3afec990-7323-4304-997d-cb2ccfd9a39b" (UID: "3afec990-7323-4304-997d-cb2ccfd9a39b"). InnerVolumeSpecName "kube-api-access-5w84t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.501532 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.501570 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w84t\" (UniqueName: \"kubernetes.io/projected/3afec990-7323-4304-997d-cb2ccfd9a39b-kube-api-access-5w84t\") on node \"crc\" DevicePath \"\"" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.519822 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3afec990-7323-4304-997d-cb2ccfd9a39b" (UID: "3afec990-7323-4304-997d-cb2ccfd9a39b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.611319 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3afec990-7323-4304-997d-cb2ccfd9a39b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.882825 4966 generic.go:334] "Generic (PLEG): container finished" podID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerID="94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c" exitCode=0 Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.882894 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q92zd" event={"ID":"3afec990-7323-4304-997d-cb2ccfd9a39b","Type":"ContainerDied","Data":"94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c"} Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.882918 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q92zd" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.882926 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q92zd" event={"ID":"3afec990-7323-4304-997d-cb2ccfd9a39b","Type":"ContainerDied","Data":"328fa47fc0d8f122bab40a91500cd7922a1cb33e51c20fd8eb9963c0c84b3d90"} Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.882946 4966 scope.go:117] "RemoveContainer" containerID="94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.911141 4966 scope.go:117] "RemoveContainer" containerID="45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.923841 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q92zd"] Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.932195 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q92zd"] Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.943055 4966 scope.go:117] "RemoveContainer" containerID="73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.986405 4966 scope.go:117] "RemoveContainer" containerID="94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c" Dec 17 10:39:37 crc kubenswrapper[4966]: E1217 10:39:37.987012 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c\": container with ID starting with 94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c not found: ID does not exist" containerID="94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.987254 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c"} err="failed to get container status \"94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c\": rpc error: code = NotFound desc = could not find container \"94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c\": container with ID starting with 94ef511def8ef3f036221374dad797a01e00c62ddb667b02e035a2a224fcf13c not found: ID does not exist" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.987292 4966 scope.go:117] "RemoveContainer" containerID="45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6" Dec 17 10:39:37 crc kubenswrapper[4966]: E1217 10:39:37.988008 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6\": container with ID starting with 45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6 not found: ID does not exist" containerID="45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.988079 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6"} err="failed to get container status \"45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6\": rpc error: code = NotFound desc = could not find container \"45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6\": container with ID starting with 45537c0a5a2d7c85f807aebde5858c3bee0d215d2501b332aa7fed2f5aa09cd6 not found: ID does not exist" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.988121 4966 scope.go:117] "RemoveContainer" containerID="73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104" Dec 17 10:39:37 crc kubenswrapper[4966]: E1217 10:39:37.988599 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104\": container with ID starting with 73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104 not found: ID does not exist" containerID="73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104" Dec 17 10:39:37 crc kubenswrapper[4966]: I1217 10:39:37.988661 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104"} err="failed to get container status \"73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104\": rpc error: code = NotFound desc = could not find container \"73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104\": container with ID starting with 73794dd0ac3b2689087ee3b896eaac34ff1eacb0b986f523dc0b0707a5f54104 not found: ID does not exist" Dec 17 10:39:38 crc kubenswrapper[4966]: I1217 10:39:38.842486 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" path="/var/lib/kubelet/pods/3afec990-7323-4304-997d-cb2ccfd9a39b/volumes" Dec 17 10:39:40 crc kubenswrapper[4966]: I1217 10:39:40.833420 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:39:40 crc kubenswrapper[4966]: E1217 10:39:40.834196 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:39:51 crc kubenswrapper[4966]: I1217 10:39:51.831072 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:39:51 crc kubenswrapper[4966]: E1217 10:39:51.832017 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:40:05 crc kubenswrapper[4966]: I1217 10:40:05.831798 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:40:05 crc kubenswrapper[4966]: E1217 10:40:05.832391 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:40:16 crc kubenswrapper[4966]: I1217 10:40:16.831553 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:40:16 crc kubenswrapper[4966]: E1217 10:40:16.832741 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:40:29 crc kubenswrapper[4966]: I1217 10:40:29.831516 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:40:29 crc kubenswrapper[4966]: E1217 10:40:29.833490 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:40:44 crc kubenswrapper[4966]: I1217 10:40:44.830597 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:40:44 crc kubenswrapper[4966]: E1217 10:40:44.832491 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:40:56 crc kubenswrapper[4966]: I1217 10:40:56.830036 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:40:56 crc kubenswrapper[4966]: E1217 10:40:56.830803 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.760167 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g5fz9"] Dec 17 10:41:05 crc kubenswrapper[4966]: E1217 10:41:05.761258 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerName="registry-server" Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.761284 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerName="registry-server" Dec 17 10:41:05 crc kubenswrapper[4966]: E1217 10:41:05.761326 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerName="extract-utilities" Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.761337 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerName="extract-utilities" Dec 17 10:41:05 crc kubenswrapper[4966]: E1217 10:41:05.761379 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerName="extract-content" Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.761387 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerName="extract-content" Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.762500 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="3afec990-7323-4304-997d-cb2ccfd9a39b" containerName="registry-server" Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.764666 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.777388 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g5fz9"] Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.912916 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npp44\" (UniqueName: \"kubernetes.io/projected/8d35720e-452c-4d5c-8f0b-af8895bebf1b-kube-api-access-npp44\") pod \"community-operators-g5fz9\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.913043 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-catalog-content\") pod \"community-operators-g5fz9\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:05 crc kubenswrapper[4966]: I1217 10:41:05.913190 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-utilities\") pod \"community-operators-g5fz9\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:06 crc kubenswrapper[4966]: I1217 10:41:06.015173 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npp44\" (UniqueName: \"kubernetes.io/projected/8d35720e-452c-4d5c-8f0b-af8895bebf1b-kube-api-access-npp44\") pod \"community-operators-g5fz9\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:06 crc kubenswrapper[4966]: I1217 10:41:06.015265 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-catalog-content\") pod \"community-operators-g5fz9\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:06 crc kubenswrapper[4966]: I1217 10:41:06.015331 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-utilities\") pod \"community-operators-g5fz9\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:06 crc kubenswrapper[4966]: I1217 10:41:06.015798 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-utilities\") pod \"community-operators-g5fz9\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:06 crc kubenswrapper[4966]: I1217 10:41:06.016500 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-catalog-content\") pod \"community-operators-g5fz9\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:06 crc kubenswrapper[4966]: I1217 10:41:06.039932 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npp44\" (UniqueName: \"kubernetes.io/projected/8d35720e-452c-4d5c-8f0b-af8895bebf1b-kube-api-access-npp44\") pod \"community-operators-g5fz9\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:06 crc kubenswrapper[4966]: I1217 10:41:06.098039 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:06 crc kubenswrapper[4966]: I1217 10:41:06.588698 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g5fz9"] Dec 17 10:41:07 crc kubenswrapper[4966]: I1217 10:41:07.418371 4966 generic.go:334] "Generic (PLEG): container finished" podID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerID="0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643" exitCode=0 Dec 17 10:41:07 crc kubenswrapper[4966]: I1217 10:41:07.418461 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5fz9" event={"ID":"8d35720e-452c-4d5c-8f0b-af8895bebf1b","Type":"ContainerDied","Data":"0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643"} Dec 17 10:41:07 crc kubenswrapper[4966]: I1217 10:41:07.418743 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5fz9" event={"ID":"8d35720e-452c-4d5c-8f0b-af8895bebf1b","Type":"ContainerStarted","Data":"9593ab041f224787ad54d9ac90bde8f44f00e49179940402a2aca935cc7a2f80"} Dec 17 10:41:07 crc kubenswrapper[4966]: I1217 10:41:07.421415 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 10:41:08 crc kubenswrapper[4966]: I1217 10:41:08.835272 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:41:08 crc kubenswrapper[4966]: E1217 10:41:08.835675 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:41:09 crc kubenswrapper[4966]: I1217 10:41:09.441390 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5fz9" event={"ID":"8d35720e-452c-4d5c-8f0b-af8895bebf1b","Type":"ContainerStarted","Data":"272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e"} Dec 17 10:41:10 crc kubenswrapper[4966]: I1217 10:41:10.454113 4966 generic.go:334] "Generic (PLEG): container finished" podID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerID="272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e" exitCode=0 Dec 17 10:41:10 crc kubenswrapper[4966]: I1217 10:41:10.454233 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5fz9" event={"ID":"8d35720e-452c-4d5c-8f0b-af8895bebf1b","Type":"ContainerDied","Data":"272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e"} Dec 17 10:41:11 crc kubenswrapper[4966]: I1217 10:41:11.465723 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5fz9" event={"ID":"8d35720e-452c-4d5c-8f0b-af8895bebf1b","Type":"ContainerStarted","Data":"000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039"} Dec 17 10:41:11 crc kubenswrapper[4966]: I1217 10:41:11.492572 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g5fz9" podStartSLOduration=2.940657054 podStartE2EDuration="6.492552778s" podCreationTimestamp="2025-12-17 10:41:05 +0000 UTC" firstStartedPulling="2025-12-17 10:41:07.421161057 +0000 UTC m=+8402.966230999" lastFinishedPulling="2025-12-17 10:41:10.973056771 +0000 UTC m=+8406.518126723" observedRunningTime="2025-12-17 10:41:11.482541295 +0000 UTC m=+8407.027611257" watchObservedRunningTime="2025-12-17 10:41:11.492552778 +0000 UTC m=+8407.037622720" Dec 17 10:41:16 crc kubenswrapper[4966]: I1217 10:41:16.099216 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:16 crc kubenswrapper[4966]: I1217 10:41:16.099631 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:16 crc kubenswrapper[4966]: I1217 10:41:16.171051 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:16 crc kubenswrapper[4966]: I1217 10:41:16.604355 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:16 crc kubenswrapper[4966]: I1217 10:41:16.658707 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g5fz9"] Dec 17 10:41:18 crc kubenswrapper[4966]: I1217 10:41:18.563276 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g5fz9" podUID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerName="registry-server" containerID="cri-o://000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039" gracePeriod=2 Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.087374 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.206891 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npp44\" (UniqueName: \"kubernetes.io/projected/8d35720e-452c-4d5c-8f0b-af8895bebf1b-kube-api-access-npp44\") pod \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.207422 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-utilities\") pod \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.207763 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-catalog-content\") pod \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\" (UID: \"8d35720e-452c-4d5c-8f0b-af8895bebf1b\") " Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.208248 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-utilities" (OuterVolumeSpecName: "utilities") pod "8d35720e-452c-4d5c-8f0b-af8895bebf1b" (UID: "8d35720e-452c-4d5c-8f0b-af8895bebf1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.208835 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.214465 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d35720e-452c-4d5c-8f0b-af8895bebf1b-kube-api-access-npp44" (OuterVolumeSpecName: "kube-api-access-npp44") pod "8d35720e-452c-4d5c-8f0b-af8895bebf1b" (UID: "8d35720e-452c-4d5c-8f0b-af8895bebf1b"). InnerVolumeSpecName "kube-api-access-npp44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.289315 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d35720e-452c-4d5c-8f0b-af8895bebf1b" (UID: "8d35720e-452c-4d5c-8f0b-af8895bebf1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.310795 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npp44\" (UniqueName: \"kubernetes.io/projected/8d35720e-452c-4d5c-8f0b-af8895bebf1b-kube-api-access-npp44\") on node \"crc\" DevicePath \"\"" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.310841 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d35720e-452c-4d5c-8f0b-af8895bebf1b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.573982 4966 generic.go:334] "Generic (PLEG): container finished" podID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerID="000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039" exitCode=0 Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.574028 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g5fz9" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.574029 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5fz9" event={"ID":"8d35720e-452c-4d5c-8f0b-af8895bebf1b","Type":"ContainerDied","Data":"000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039"} Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.574083 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g5fz9" event={"ID":"8d35720e-452c-4d5c-8f0b-af8895bebf1b","Type":"ContainerDied","Data":"9593ab041f224787ad54d9ac90bde8f44f00e49179940402a2aca935cc7a2f80"} Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.574104 4966 scope.go:117] "RemoveContainer" containerID="000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.600860 4966 scope.go:117] "RemoveContainer" containerID="272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.615960 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g5fz9"] Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.631531 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g5fz9"] Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.647187 4966 scope.go:117] "RemoveContainer" containerID="0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.699296 4966 scope.go:117] "RemoveContainer" containerID="000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039" Dec 17 10:41:19 crc kubenswrapper[4966]: E1217 10:41:19.700128 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039\": container with ID starting with 000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039 not found: ID does not exist" containerID="000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.700176 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039"} err="failed to get container status \"000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039\": rpc error: code = NotFound desc = could not find container \"000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039\": container with ID starting with 000ae0f06d3a4ccba94d913636d0c0b672441bcc7c35fe9fa0748877270cb039 not found: ID does not exist" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.700199 4966 scope.go:117] "RemoveContainer" containerID="272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e" Dec 17 10:41:19 crc kubenswrapper[4966]: E1217 10:41:19.700459 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e\": container with ID starting with 272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e not found: ID does not exist" containerID="272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.700508 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e"} err="failed to get container status \"272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e\": rpc error: code = NotFound desc = could not find container \"272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e\": container with ID starting with 272a58179181872218a29a37f40590f8e457f967a85e6d7e0a49e8447b5e652e not found: ID does not exist" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.700525 4966 scope.go:117] "RemoveContainer" containerID="0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643" Dec 17 10:41:19 crc kubenswrapper[4966]: E1217 10:41:19.700972 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643\": container with ID starting with 0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643 not found: ID does not exist" containerID="0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643" Dec 17 10:41:19 crc kubenswrapper[4966]: I1217 10:41:19.700998 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643"} err="failed to get container status \"0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643\": rpc error: code = NotFound desc = could not find container \"0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643\": container with ID starting with 0bfb2e12ebf4c621437eeb890b94ce18dc7f33561e4e16b0c5b4bbbd80161643 not found: ID does not exist" Dec 17 10:41:20 crc kubenswrapper[4966]: I1217 10:41:20.854349 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" path="/var/lib/kubelet/pods/8d35720e-452c-4d5c-8f0b-af8895bebf1b/volumes" Dec 17 10:41:21 crc kubenswrapper[4966]: I1217 10:41:21.831352 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:41:21 crc kubenswrapper[4966]: E1217 10:41:21.832045 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:41:35 crc kubenswrapper[4966]: I1217 10:41:35.830060 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:41:35 crc kubenswrapper[4966]: E1217 10:41:35.830664 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:41:50 crc kubenswrapper[4966]: I1217 10:41:50.831183 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:41:50 crc kubenswrapper[4966]: E1217 10:41:50.832164 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:42:01 crc kubenswrapper[4966]: I1217 10:42:01.830215 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:42:01 crc kubenswrapper[4966]: E1217 10:42:01.830979 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:42:13 crc kubenswrapper[4966]: I1217 10:42:13.830336 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:42:13 crc kubenswrapper[4966]: E1217 10:42:13.831270 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:42:27 crc kubenswrapper[4966]: I1217 10:42:27.831041 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:42:29 crc kubenswrapper[4966]: I1217 10:42:29.203776 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"be0da4b76c123744b21b8ebc207645cfead54d433f111b12928f641bbcb96caf"} Dec 17 10:44:46 crc kubenswrapper[4966]: I1217 10:44:46.808269 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:44:46 crc kubenswrapper[4966]: I1217 10:44:46.808976 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.263244 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6"] Dec 17 10:45:00 crc kubenswrapper[4966]: E1217 10:45:00.264228 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerName="extract-utilities" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.264247 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerName="extract-utilities" Dec 17 10:45:00 crc kubenswrapper[4966]: E1217 10:45:00.264292 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerName="registry-server" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.264302 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerName="registry-server" Dec 17 10:45:00 crc kubenswrapper[4966]: E1217 10:45:00.264320 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerName="extract-content" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.264327 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerName="extract-content" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.264577 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d35720e-452c-4d5c-8f0b-af8895bebf1b" containerName="registry-server" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.265622 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.272308 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.274691 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.284198 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6"] Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.399428 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqntf\" (UniqueName: \"kubernetes.io/projected/cb863ced-60bf-4625-a094-73cbe7b63212-kube-api-access-hqntf\") pod \"collect-profiles-29432805-sznd6\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.399486 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb863ced-60bf-4625-a094-73cbe7b63212-config-volume\") pod \"collect-profiles-29432805-sznd6\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.399583 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb863ced-60bf-4625-a094-73cbe7b63212-secret-volume\") pod \"collect-profiles-29432805-sznd6\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.501140 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqntf\" (UniqueName: \"kubernetes.io/projected/cb863ced-60bf-4625-a094-73cbe7b63212-kube-api-access-hqntf\") pod \"collect-profiles-29432805-sznd6\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.501207 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb863ced-60bf-4625-a094-73cbe7b63212-config-volume\") pod \"collect-profiles-29432805-sznd6\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.501270 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb863ced-60bf-4625-a094-73cbe7b63212-secret-volume\") pod \"collect-profiles-29432805-sznd6\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.502832 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb863ced-60bf-4625-a094-73cbe7b63212-config-volume\") pod \"collect-profiles-29432805-sznd6\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.513039 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb863ced-60bf-4625-a094-73cbe7b63212-secret-volume\") pod \"collect-profiles-29432805-sznd6\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.517427 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqntf\" (UniqueName: \"kubernetes.io/projected/cb863ced-60bf-4625-a094-73cbe7b63212-kube-api-access-hqntf\") pod \"collect-profiles-29432805-sznd6\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:00 crc kubenswrapper[4966]: I1217 10:45:00.588624 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:01 crc kubenswrapper[4966]: I1217 10:45:01.095644 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6"] Dec 17 10:45:01 crc kubenswrapper[4966]: I1217 10:45:01.631539 4966 generic.go:334] "Generic (PLEG): container finished" podID="cb863ced-60bf-4625-a094-73cbe7b63212" containerID="de1fc454500531718af945411950ad5f96efe81540c596f0091a695787e0c52e" exitCode=0 Dec 17 10:45:01 crc kubenswrapper[4966]: I1217 10:45:01.631793 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" event={"ID":"cb863ced-60bf-4625-a094-73cbe7b63212","Type":"ContainerDied","Data":"de1fc454500531718af945411950ad5f96efe81540c596f0091a695787e0c52e"} Dec 17 10:45:01 crc kubenswrapper[4966]: I1217 10:45:01.631817 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" event={"ID":"cb863ced-60bf-4625-a094-73cbe7b63212","Type":"ContainerStarted","Data":"203bcafb6d1243c9865acb8cbae96b1d45663e7fb5d919593ff9fd10cd0d6e8d"} Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.024177 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.151200 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb863ced-60bf-4625-a094-73cbe7b63212-config-volume\") pod \"cb863ced-60bf-4625-a094-73cbe7b63212\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.151445 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqntf\" (UniqueName: \"kubernetes.io/projected/cb863ced-60bf-4625-a094-73cbe7b63212-kube-api-access-hqntf\") pod \"cb863ced-60bf-4625-a094-73cbe7b63212\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.151489 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb863ced-60bf-4625-a094-73cbe7b63212-secret-volume\") pod \"cb863ced-60bf-4625-a094-73cbe7b63212\" (UID: \"cb863ced-60bf-4625-a094-73cbe7b63212\") " Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.152447 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb863ced-60bf-4625-a094-73cbe7b63212-config-volume" (OuterVolumeSpecName: "config-volume") pod "cb863ced-60bf-4625-a094-73cbe7b63212" (UID: "cb863ced-60bf-4625-a094-73cbe7b63212"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.152787 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cb863ced-60bf-4625-a094-73cbe7b63212-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.157664 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb863ced-60bf-4625-a094-73cbe7b63212-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cb863ced-60bf-4625-a094-73cbe7b63212" (UID: "cb863ced-60bf-4625-a094-73cbe7b63212"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.158441 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb863ced-60bf-4625-a094-73cbe7b63212-kube-api-access-hqntf" (OuterVolumeSpecName: "kube-api-access-hqntf") pod "cb863ced-60bf-4625-a094-73cbe7b63212" (UID: "cb863ced-60bf-4625-a094-73cbe7b63212"). InnerVolumeSpecName "kube-api-access-hqntf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.274355 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqntf\" (UniqueName: \"kubernetes.io/projected/cb863ced-60bf-4625-a094-73cbe7b63212-kube-api-access-hqntf\") on node \"crc\" DevicePath \"\"" Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.274383 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cb863ced-60bf-4625-a094-73cbe7b63212-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.648547 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" event={"ID":"cb863ced-60bf-4625-a094-73cbe7b63212","Type":"ContainerDied","Data":"203bcafb6d1243c9865acb8cbae96b1d45663e7fb5d919593ff9fd10cd0d6e8d"} Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.648591 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="203bcafb6d1243c9865acb8cbae96b1d45663e7fb5d919593ff9fd10cd0d6e8d" Dec 17 10:45:03 crc kubenswrapper[4966]: I1217 10:45:03.648668 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432805-sznd6" Dec 17 10:45:04 crc kubenswrapper[4966]: I1217 10:45:04.124638 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk"] Dec 17 10:45:04 crc kubenswrapper[4966]: I1217 10:45:04.134497 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432760-6pzqk"] Dec 17 10:45:04 crc kubenswrapper[4966]: I1217 10:45:04.844089 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b" path="/var/lib/kubelet/pods/a27a418b-f81c-4546-9e7f-f0c4a8d1ae0b/volumes" Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.807829 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.808326 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.895592 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wkrbr"] Dec 17 10:45:16 crc kubenswrapper[4966]: E1217 10:45:16.896119 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb863ced-60bf-4625-a094-73cbe7b63212" containerName="collect-profiles" Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.896143 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb863ced-60bf-4625-a094-73cbe7b63212" containerName="collect-profiles" Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.896447 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb863ced-60bf-4625-a094-73cbe7b63212" containerName="collect-profiles" Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.898233 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.913719 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wkrbr"] Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.950253 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/571a0368-eb5d-40ea-a013-add01af831d9-utilities\") pod \"redhat-operators-wkrbr\" (UID: \"571a0368-eb5d-40ea-a013-add01af831d9\") " pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.950411 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/571a0368-eb5d-40ea-a013-add01af831d9-catalog-content\") pod \"redhat-operators-wkrbr\" (UID: \"571a0368-eb5d-40ea-a013-add01af831d9\") " pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:16 crc kubenswrapper[4966]: I1217 10:45:16.950466 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs7v6\" (UniqueName: \"kubernetes.io/projected/571a0368-eb5d-40ea-a013-add01af831d9-kube-api-access-zs7v6\") pod \"redhat-operators-wkrbr\" (UID: \"571a0368-eb5d-40ea-a013-add01af831d9\") " pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:17 crc kubenswrapper[4966]: I1217 10:45:17.051803 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/571a0368-eb5d-40ea-a013-add01af831d9-utilities\") pod \"redhat-operators-wkrbr\" (UID: \"571a0368-eb5d-40ea-a013-add01af831d9\") " pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:17 crc kubenswrapper[4966]: I1217 10:45:17.051915 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/571a0368-eb5d-40ea-a013-add01af831d9-catalog-content\") pod \"redhat-operators-wkrbr\" (UID: \"571a0368-eb5d-40ea-a013-add01af831d9\") " pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:17 crc kubenswrapper[4966]: I1217 10:45:17.051940 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs7v6\" (UniqueName: \"kubernetes.io/projected/571a0368-eb5d-40ea-a013-add01af831d9-kube-api-access-zs7v6\") pod \"redhat-operators-wkrbr\" (UID: \"571a0368-eb5d-40ea-a013-add01af831d9\") " pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:17 crc kubenswrapper[4966]: I1217 10:45:17.052655 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/571a0368-eb5d-40ea-a013-add01af831d9-utilities\") pod \"redhat-operators-wkrbr\" (UID: \"571a0368-eb5d-40ea-a013-add01af831d9\") " pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:17 crc kubenswrapper[4966]: I1217 10:45:17.052778 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/571a0368-eb5d-40ea-a013-add01af831d9-catalog-content\") pod \"redhat-operators-wkrbr\" (UID: \"571a0368-eb5d-40ea-a013-add01af831d9\") " pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:17 crc kubenswrapper[4966]: I1217 10:45:17.073017 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs7v6\" (UniqueName: \"kubernetes.io/projected/571a0368-eb5d-40ea-a013-add01af831d9-kube-api-access-zs7v6\") pod \"redhat-operators-wkrbr\" (UID: \"571a0368-eb5d-40ea-a013-add01af831d9\") " pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:17 crc kubenswrapper[4966]: I1217 10:45:17.231386 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:17 crc kubenswrapper[4966]: I1217 10:45:17.725950 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wkrbr"] Dec 17 10:45:17 crc kubenswrapper[4966]: W1217 10:45:17.728189 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod571a0368_eb5d_40ea_a013_add01af831d9.slice/crio-4ef4b6eee12ee98aacdd829c735ef7d4f7c9e2e0f8648dc4038d7143ca3e84b7 WatchSource:0}: Error finding container 4ef4b6eee12ee98aacdd829c735ef7d4f7c9e2e0f8648dc4038d7143ca3e84b7: Status 404 returned error can't find the container with id 4ef4b6eee12ee98aacdd829c735ef7d4f7c9e2e0f8648dc4038d7143ca3e84b7 Dec 17 10:45:17 crc kubenswrapper[4966]: I1217 10:45:17.793813 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wkrbr" event={"ID":"571a0368-eb5d-40ea-a013-add01af831d9","Type":"ContainerStarted","Data":"4ef4b6eee12ee98aacdd829c735ef7d4f7c9e2e0f8648dc4038d7143ca3e84b7"} Dec 17 10:45:18 crc kubenswrapper[4966]: I1217 10:45:18.802844 4966 generic.go:334] "Generic (PLEG): container finished" podID="571a0368-eb5d-40ea-a013-add01af831d9" containerID="ada6d72a3fbc042ce80e56e0b694ba02f3150609efa3a3fe7664cb651d651dd0" exitCode=0 Dec 17 10:45:18 crc kubenswrapper[4966]: I1217 10:45:18.803150 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wkrbr" event={"ID":"571a0368-eb5d-40ea-a013-add01af831d9","Type":"ContainerDied","Data":"ada6d72a3fbc042ce80e56e0b694ba02f3150609efa3a3fe7664cb651d651dd0"} Dec 17 10:45:31 crc kubenswrapper[4966]: I1217 10:45:31.930914 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wkrbr" event={"ID":"571a0368-eb5d-40ea-a013-add01af831d9","Type":"ContainerStarted","Data":"2a379b2c4e71f183b7be739e2ae41644eee9b1992da37363cd6c20f29db4e164"} Dec 17 10:45:35 crc kubenswrapper[4966]: I1217 10:45:35.971119 4966 generic.go:334] "Generic (PLEG): container finished" podID="571a0368-eb5d-40ea-a013-add01af831d9" containerID="2a379b2c4e71f183b7be739e2ae41644eee9b1992da37363cd6c20f29db4e164" exitCode=0 Dec 17 10:45:35 crc kubenswrapper[4966]: I1217 10:45:35.971223 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wkrbr" event={"ID":"571a0368-eb5d-40ea-a013-add01af831d9","Type":"ContainerDied","Data":"2a379b2c4e71f183b7be739e2ae41644eee9b1992da37363cd6c20f29db4e164"} Dec 17 10:45:36 crc kubenswrapper[4966]: I1217 10:45:36.987601 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wkrbr" event={"ID":"571a0368-eb5d-40ea-a013-add01af831d9","Type":"ContainerStarted","Data":"e1f5dc8e18605aa019e02bae1ddd14c43e9ece1eee28f083f4946438dbc43bd1"} Dec 17 10:45:37 crc kubenswrapper[4966]: I1217 10:45:37.015445 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wkrbr" podStartSLOduration=3.192580933 podStartE2EDuration="21.015427607s" podCreationTimestamp="2025-12-17 10:45:16 +0000 UTC" firstStartedPulling="2025-12-17 10:45:18.805137694 +0000 UTC m=+8654.350207636" lastFinishedPulling="2025-12-17 10:45:36.627984368 +0000 UTC m=+8672.173054310" observedRunningTime="2025-12-17 10:45:37.010493242 +0000 UTC m=+8672.555563184" watchObservedRunningTime="2025-12-17 10:45:37.015427607 +0000 UTC m=+8672.560497549" Dec 17 10:45:37 crc kubenswrapper[4966]: I1217 10:45:37.231986 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:37 crc kubenswrapper[4966]: I1217 10:45:37.232038 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:38 crc kubenswrapper[4966]: I1217 10:45:38.277219 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wkrbr" podUID="571a0368-eb5d-40ea-a013-add01af831d9" containerName="registry-server" probeResult="failure" output=< Dec 17 10:45:38 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:45:38 crc kubenswrapper[4966]: > Dec 17 10:45:46 crc kubenswrapper[4966]: I1217 10:45:46.807949 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:45:46 crc kubenswrapper[4966]: I1217 10:45:46.808538 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:45:46 crc kubenswrapper[4966]: I1217 10:45:46.808592 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:45:46 crc kubenswrapper[4966]: I1217 10:45:46.809475 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be0da4b76c123744b21b8ebc207645cfead54d433f111b12928f641bbcb96caf"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:45:46 crc kubenswrapper[4966]: I1217 10:45:46.809526 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://be0da4b76c123744b21b8ebc207645cfead54d433f111b12928f641bbcb96caf" gracePeriod=600 Dec 17 10:45:47 crc kubenswrapper[4966]: I1217 10:45:47.079434 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="be0da4b76c123744b21b8ebc207645cfead54d433f111b12928f641bbcb96caf" exitCode=0 Dec 17 10:45:47 crc kubenswrapper[4966]: I1217 10:45:47.079790 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"be0da4b76c123744b21b8ebc207645cfead54d433f111b12928f641bbcb96caf"} Dec 17 10:45:47 crc kubenswrapper[4966]: I1217 10:45:47.079831 4966 scope.go:117] "RemoveContainer" containerID="31c79505ee98fb7a42616701048ad2236a17357b78a7c35d305687301cd4b5ef" Dec 17 10:45:48 crc kubenswrapper[4966]: I1217 10:45:48.090175 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5"} Dec 17 10:45:48 crc kubenswrapper[4966]: I1217 10:45:48.282042 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wkrbr" podUID="571a0368-eb5d-40ea-a013-add01af831d9" containerName="registry-server" probeResult="failure" output=< Dec 17 10:45:48 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:45:48 crc kubenswrapper[4966]: > Dec 17 10:45:57 crc kubenswrapper[4966]: I1217 10:45:57.302668 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:57 crc kubenswrapper[4966]: I1217 10:45:57.365078 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wkrbr" Dec 17 10:45:57 crc kubenswrapper[4966]: I1217 10:45:57.445268 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wkrbr"] Dec 17 10:45:57 crc kubenswrapper[4966]: I1217 10:45:57.549584 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c9l6f"] Dec 17 10:45:57 crc kubenswrapper[4966]: I1217 10:45:57.550068 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c9l6f" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" containerName="registry-server" containerID="cri-o://9af75d3f8f4a710b951f69ce7a34f48209b303ed6a34b72a208f64e3bb82577b" gracePeriod=2 Dec 17 10:45:57 crc kubenswrapper[4966]: I1217 10:45:57.749226 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-db25g"] Dec 17 10:45:57 crc kubenswrapper[4966]: I1217 10:45:57.749719 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-db25g" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerName="registry-server" containerID="cri-o://09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f" gracePeriod=2 Dec 17 10:45:57 crc kubenswrapper[4966]: I1217 10:45:57.950013 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gvr6s"] Dec 17 10:45:57 crc kubenswrapper[4966]: I1217 10:45:57.950345 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gvr6s" podUID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerName="registry-server" containerID="cri-o://8a03fe11e7a3d7a669a8a140e8147a154fad325a66c3d030d96299652f39c054" gracePeriod=2 Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.167648 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pv2kq"] Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.167992 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pv2kq" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="registry-server" containerID="cri-o://1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790" gracePeriod=2 Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.233975 4966 generic.go:334] "Generic (PLEG): container finished" podID="58b62b06-5dd1-46d2-a69b-767595012e59" containerID="9af75d3f8f4a710b951f69ce7a34f48209b303ed6a34b72a208f64e3bb82577b" exitCode=0 Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.234037 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9l6f" event={"ID":"58b62b06-5dd1-46d2-a69b-767595012e59","Type":"ContainerDied","Data":"9af75d3f8f4a710b951f69ce7a34f48209b303ed6a34b72a208f64e3bb82577b"} Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.251978 4966 generic.go:334] "Generic (PLEG): container finished" podID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerID="8a03fe11e7a3d7a669a8a140e8147a154fad325a66c3d030d96299652f39c054" exitCode=0 Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.252526 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvr6s" event={"ID":"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c","Type":"ContainerDied","Data":"8a03fe11e7a3d7a669a8a140e8147a154fad325a66c3d030d96299652f39c054"} Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.474410 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.631390 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-utilities\") pod \"58b62b06-5dd1-46d2-a69b-767595012e59\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.631569 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-catalog-content\") pod \"58b62b06-5dd1-46d2-a69b-767595012e59\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.631606 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbj8g\" (UniqueName: \"kubernetes.io/projected/58b62b06-5dd1-46d2-a69b-767595012e59-kube-api-access-mbj8g\") pod \"58b62b06-5dd1-46d2-a69b-767595012e59\" (UID: \"58b62b06-5dd1-46d2-a69b-767595012e59\") " Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.632323 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-utilities" (OuterVolumeSpecName: "utilities") pod "58b62b06-5dd1-46d2-a69b-767595012e59" (UID: "58b62b06-5dd1-46d2-a69b-767595012e59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.634253 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.665822 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58b62b06-5dd1-46d2-a69b-767595012e59-kube-api-access-mbj8g" (OuterVolumeSpecName: "kube-api-access-mbj8g") pod "58b62b06-5dd1-46d2-a69b-767595012e59" (UID: "58b62b06-5dd1-46d2-a69b-767595012e59"). InnerVolumeSpecName "kube-api-access-mbj8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.747694 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbj8g\" (UniqueName: \"kubernetes.io/projected/58b62b06-5dd1-46d2-a69b-767595012e59-kube-api-access-mbj8g\") on node \"crc\" DevicePath \"\"" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.777346 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58b62b06-5dd1-46d2-a69b-767595012e59" (UID: "58b62b06-5dd1-46d2-a69b-767595012e59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.823694 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.849751 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58b62b06-5dd1-46d2-a69b-767595012e59-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.950712 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-catalog-content\") pod \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.951227 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-utilities\") pod \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.951388 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7chsc\" (UniqueName: \"kubernetes.io/projected/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-kube-api-access-7chsc\") pod \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\" (UID: \"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c\") " Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.951535 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-utilities" (OuterVolumeSpecName: "utilities") pod "9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" (UID: "9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.952908 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:45:58 crc kubenswrapper[4966]: I1217 10:45:58.956554 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-kube-api-access-7chsc" (OuterVolumeSpecName: "kube-api-access-7chsc") pod "9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" (UID: "9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c"). InnerVolumeSpecName "kube-api-access-7chsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.025227 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" (UID: "9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.054402 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7chsc\" (UniqueName: \"kubernetes.io/projected/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-kube-api-access-7chsc\") on node \"crc\" DevicePath \"\"" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.054437 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.262377 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvr6s" event={"ID":"9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c","Type":"ContainerDied","Data":"2fd4c5e1c7953e3744400d7c08094799f290ce4ed569f7ee8c672d09febcfcdb"} Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.262432 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gvr6s" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.262444 4966 scope.go:117] "RemoveContainer" containerID="8a03fe11e7a3d7a669a8a140e8147a154fad325a66c3d030d96299652f39c054" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.265759 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9l6f" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.266252 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9l6f" event={"ID":"58b62b06-5dd1-46d2-a69b-767595012e59","Type":"ContainerDied","Data":"a4380601fd215276280c396b9fca9e1127bb4bd273693bf9a046c15515366922"} Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.293080 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c9l6f"] Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.302037 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c9l6f"] Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.320096 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gvr6s"] Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.328382 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gvr6s"] Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.376833 4966 scope.go:117] "RemoveContainer" containerID="0277cb0975789b6b645e1f55ebc25cf546ac333381c0d31cc605fb271772ece3" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.404254 4966 scope.go:117] "RemoveContainer" containerID="9fa20c5d41bb583227a1ef0070db1556f83021a47856052d810bb227051d36f8" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.479557 4966 scope.go:117] "RemoveContainer" containerID="9af75d3f8f4a710b951f69ce7a34f48209b303ed6a34b72a208f64e3bb82577b" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.503652 4966 scope.go:117] "RemoveContainer" containerID="c18b05a228a591a133668047e295d663459afc446fb2aa58c0c386ef764235dc" Dec 17 10:45:59 crc kubenswrapper[4966]: I1217 10:45:59.588332 4966 scope.go:117] "RemoveContainer" containerID="7f313097a3fcfdd4729bd70258b4c859b6520cb5075269aa7b390f9d1db418ac" Dec 17 10:46:05 crc kubenswrapper[4966]: I1217 10:46:02.048097 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-84b495f78-tgd62" podUID="0be98506-68e8-438f-a46b-2225dd61b871" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.79:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 17 10:46:05 crc kubenswrapper[4966]: I1217 10:46:04.213723 4966 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-j476q container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 17 10:46:05 crc kubenswrapper[4966]: I1217 10:46:04.213796 4966 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-j476q" podUID="f25b1b4a-7676-4f7c-847a-ee29217bb19d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 17 10:46:05 crc kubenswrapper[4966]: E1217 10:46:05.272732 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790 is running failed: container process not found" containerID="1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790" cmd=["grpc_health_probe","-addr=:50051"] Dec 17 10:46:05 crc kubenswrapper[4966]: I1217 10:46:05.279838 4966 scope.go:117] "RemoveContainer" containerID="6756d56218b2294c5a2fe45c00185ff724671f2d1305b390694d00893dd34b26" Dec 17 10:46:05 crc kubenswrapper[4966]: E1217 10:46:05.313466 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790 is running failed: container process not found" containerID="1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790" cmd=["grpc_health_probe","-addr=:50051"] Dec 17 10:46:05 crc kubenswrapper[4966]: E1217 10:46:05.355760 4966 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790 is running failed: container process not found" containerID="1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790" cmd=["grpc_health_probe","-addr=:50051"] Dec 17 10:46:05 crc kubenswrapper[4966]: E1217 10:46:05.355840 4966 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-pv2kq" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="registry-server" Dec 17 10:46:05 crc kubenswrapper[4966]: I1217 10:46:05.393569 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" path="/var/lib/kubelet/pods/58b62b06-5dd1-46d2-a69b-767595012e59/volumes" Dec 17 10:46:05 crc kubenswrapper[4966]: I1217 10:46:05.401424 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" path="/var/lib/kubelet/pods/9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c/volumes" Dec 17 10:46:05 crc kubenswrapper[4966]: E1217 10:46:05.402230 4966 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="4.573s" Dec 17 10:46:05 crc kubenswrapper[4966]: E1217 10:46:05.931014 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod982c7add_45d7_451e_8e31_3050c85eaa89.slice/crio-conmon-1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cd4b4c7_507c_4f2c_a657_b579d039441b.slice/crio-conmon-09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f.scope\": RecentStats: unable to find data in memory cache]" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.245104 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-db25g_2cd4b4c7-507c-4f2c-a657-b579d039441b/registry-server/0.log" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.246507 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db25g" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.303217 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-utilities\") pod \"2cd4b4c7-507c-4f2c-a657-b579d039441b\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.303292 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzcxz\" (UniqueName: \"kubernetes.io/projected/2cd4b4c7-507c-4f2c-a657-b579d039441b-kube-api-access-pzcxz\") pod \"2cd4b4c7-507c-4f2c-a657-b579d039441b\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.303512 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-catalog-content\") pod \"2cd4b4c7-507c-4f2c-a657-b579d039441b\" (UID: \"2cd4b4c7-507c-4f2c-a657-b579d039441b\") " Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.305112 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-utilities" (OuterVolumeSpecName: "utilities") pod "2cd4b4c7-507c-4f2c-a657-b579d039441b" (UID: "2cd4b4c7-507c-4f2c-a657-b579d039441b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.313451 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cd4b4c7-507c-4f2c-a657-b579d039441b-kube-api-access-pzcxz" (OuterVolumeSpecName: "kube-api-access-pzcxz") pod "2cd4b4c7-507c-4f2c-a657-b579d039441b" (UID: "2cd4b4c7-507c-4f2c-a657-b579d039441b"). InnerVolumeSpecName "kube-api-access-pzcxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.354335 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2cd4b4c7-507c-4f2c-a657-b579d039441b" (UID: "2cd4b4c7-507c-4f2c-a657-b579d039441b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.405714 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.405758 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cd4b4c7-507c-4f2c-a657-b579d039441b-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.405771 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzcxz\" (UniqueName: \"kubernetes.io/projected/2cd4b4c7-507c-4f2c-a657-b579d039441b-kube-api-access-pzcxz\") on node \"crc\" DevicePath \"\"" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.465582 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pv2kq_982c7add-45d7-451e-8e31-3050c85eaa89/registry-server/0.log" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.467099 4966 generic.go:334] "Generic (PLEG): container finished" podID="982c7add-45d7-451e-8e31-3050c85eaa89" containerID="1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790" exitCode=137 Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.467236 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pv2kq" event={"ID":"982c7add-45d7-451e-8e31-3050c85eaa89","Type":"ContainerDied","Data":"1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790"} Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.471705 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-db25g_2cd4b4c7-507c-4f2c-a657-b579d039441b/registry-server/0.log" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.472918 4966 generic.go:334] "Generic (PLEG): container finished" podID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerID="09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f" exitCode=137 Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.473128 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db25g" event={"ID":"2cd4b4c7-507c-4f2c-a657-b579d039441b","Type":"ContainerDied","Data":"09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f"} Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.473360 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-db25g" event={"ID":"2cd4b4c7-507c-4f2c-a657-b579d039441b","Type":"ContainerDied","Data":"b4c9c3a10f088d00e219badf94a114b66b63bdad7cc3df0ddc2ddfe3137b88aa"} Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.473500 4966 scope.go:117] "RemoveContainer" containerID="09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.473633 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-db25g" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.515062 4966 scope.go:117] "RemoveContainer" containerID="10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.519086 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-db25g"] Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.533441 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-db25g"] Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.559023 4966 scope.go:117] "RemoveContainer" containerID="50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.628158 4966 scope.go:117] "RemoveContainer" containerID="09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f" Dec 17 10:46:06 crc kubenswrapper[4966]: E1217 10:46:06.632888 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f\": container with ID starting with 09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f not found: ID does not exist" containerID="09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.632926 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f"} err="failed to get container status \"09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f\": rpc error: code = NotFound desc = could not find container \"09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f\": container with ID starting with 09acb45c0896335496f4da3bbc6ace81913707f56d6a6168b1bc97947df4e71f not found: ID does not exist" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.632952 4966 scope.go:117] "RemoveContainer" containerID="10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264" Dec 17 10:46:06 crc kubenswrapper[4966]: E1217 10:46:06.633420 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264\": container with ID starting with 10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264 not found: ID does not exist" containerID="10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.633466 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264"} err="failed to get container status \"10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264\": rpc error: code = NotFound desc = could not find container \"10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264\": container with ID starting with 10ae410bb2ecb6df6b9abf7f7c640980a3a67f418d7edcfb3acc753e63637264 not found: ID does not exist" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.633497 4966 scope.go:117] "RemoveContainer" containerID="50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca" Dec 17 10:46:06 crc kubenswrapper[4966]: E1217 10:46:06.634241 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca\": container with ID starting with 50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca not found: ID does not exist" containerID="50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.634264 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca"} err="failed to get container status \"50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca\": rpc error: code = NotFound desc = could not find container \"50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca\": container with ID starting with 50879431312f8630aafec82eaa79a626b7e9ffa4d237954d0b377d561ef8ccca not found: ID does not exist" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.847568 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" path="/var/lib/kubelet/pods/2cd4b4c7-507c-4f2c-a657-b579d039441b/volumes" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.938974 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pv2kq_982c7add-45d7-451e-8e31-3050c85eaa89/registry-server/0.log" Dec 17 10:46:06 crc kubenswrapper[4966]: I1217 10:46:06.939856 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.022529 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-catalog-content\") pod \"982c7add-45d7-451e-8e31-3050c85eaa89\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.022708 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-utilities\") pod \"982c7add-45d7-451e-8e31-3050c85eaa89\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.022779 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scnwr\" (UniqueName: \"kubernetes.io/projected/982c7add-45d7-451e-8e31-3050c85eaa89-kube-api-access-scnwr\") pod \"982c7add-45d7-451e-8e31-3050c85eaa89\" (UID: \"982c7add-45d7-451e-8e31-3050c85eaa89\") " Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.023456 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-utilities" (OuterVolumeSpecName: "utilities") pod "982c7add-45d7-451e-8e31-3050c85eaa89" (UID: "982c7add-45d7-451e-8e31-3050c85eaa89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.028093 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/982c7add-45d7-451e-8e31-3050c85eaa89-kube-api-access-scnwr" (OuterVolumeSpecName: "kube-api-access-scnwr") pod "982c7add-45d7-451e-8e31-3050c85eaa89" (UID: "982c7add-45d7-451e-8e31-3050c85eaa89"). InnerVolumeSpecName "kube-api-access-scnwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.058769 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "982c7add-45d7-451e-8e31-3050c85eaa89" (UID: "982c7add-45d7-451e-8e31-3050c85eaa89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.124818 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.125208 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scnwr\" (UniqueName: \"kubernetes.io/projected/982c7add-45d7-451e-8e31-3050c85eaa89-kube-api-access-scnwr\") on node \"crc\" DevicePath \"\"" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.125266 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/982c7add-45d7-451e-8e31-3050c85eaa89-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.495552 4966 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pv2kq_982c7add-45d7-451e-8e31-3050c85eaa89/registry-server/0.log" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.496254 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pv2kq" event={"ID":"982c7add-45d7-451e-8e31-3050c85eaa89","Type":"ContainerDied","Data":"ca7573a7d922fcf3440d801c4e853e1e8024189d0ea972a8773981b0532e1ee3"} Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.496311 4966 scope.go:117] "RemoveContainer" containerID="1416629f8d2b166894ba64d80a52f0c3d93db6069ae17c6f098cc107db54a790" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.496412 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pv2kq" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.533299 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pv2kq"] Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.543067 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pv2kq"] Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.728120 4966 scope.go:117] "RemoveContainer" containerID="9e94849486ea9a5989a2791c83a8f4292e223745832f1d41d1f51fe1da047e51" Dec 17 10:46:07 crc kubenswrapper[4966]: I1217 10:46:07.789591 4966 scope.go:117] "RemoveContainer" containerID="c683ad190b7b29e464f98c00b8cd83423cf60ace24621230effac6569365e1bf" Dec 17 10:46:08 crc kubenswrapper[4966]: I1217 10:46:08.842274 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" path="/var/lib/kubelet/pods/982c7add-45d7-451e-8e31-3050c85eaa89/volumes" Dec 17 10:48:16 crc kubenswrapper[4966]: I1217 10:48:16.807739 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:48:16 crc kubenswrapper[4966]: I1217 10:48:16.808240 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:48:46 crc kubenswrapper[4966]: I1217 10:48:46.807433 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:48:46 crc kubenswrapper[4966]: I1217 10:48:46.808108 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.800610 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9wgm4"] Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803365 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803397 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803420 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerName="extract-content" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803442 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerName="extract-content" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803452 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerName="extract-content" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803457 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerName="extract-content" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803466 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803472 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803485 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerName="extract-utilities" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803493 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerName="extract-utilities" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803525 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerName="extract-utilities" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803533 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerName="extract-utilities" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803548 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803555 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803571 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803579 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803613 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="extract-utilities" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803619 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="extract-utilities" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803631 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="extract-content" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803637 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="extract-content" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803650 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" containerName="extract-utilities" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803656 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" containerName="extract-utilities" Dec 17 10:49:06 crc kubenswrapper[4966]: E1217 10:49:06.803687 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" containerName="extract-content" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.803693 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" containerName="extract-content" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.804336 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="982c7add-45d7-451e-8e31-3050c85eaa89" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.804370 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="58b62b06-5dd1-46d2-a69b-767595012e59" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.804398 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cd4b4c7-507c-4f2c-a657-b579d039441b" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.804415 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9db233fb-f3e8-4b7d-a1ee-2f0fa974f81c" containerName="registry-server" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.806988 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.829117 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wgm4"] Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.918797 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-catalog-content\") pod \"redhat-marketplace-9wgm4\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.920095 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjfrl\" (UniqueName: \"kubernetes.io/projected/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-kube-api-access-cjfrl\") pod \"redhat-marketplace-9wgm4\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:06 crc kubenswrapper[4966]: I1217 10:49:06.920223 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-utilities\") pod \"redhat-marketplace-9wgm4\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:07 crc kubenswrapper[4966]: I1217 10:49:07.021984 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-catalog-content\") pod \"redhat-marketplace-9wgm4\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:07 crc kubenswrapper[4966]: I1217 10:49:07.022196 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjfrl\" (UniqueName: \"kubernetes.io/projected/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-kube-api-access-cjfrl\") pod \"redhat-marketplace-9wgm4\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:07 crc kubenswrapper[4966]: I1217 10:49:07.022241 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-utilities\") pod \"redhat-marketplace-9wgm4\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:07 crc kubenswrapper[4966]: I1217 10:49:07.022512 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-catalog-content\") pod \"redhat-marketplace-9wgm4\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:07 crc kubenswrapper[4966]: I1217 10:49:07.022627 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-utilities\") pod \"redhat-marketplace-9wgm4\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:07 crc kubenswrapper[4966]: I1217 10:49:07.107941 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjfrl\" (UniqueName: \"kubernetes.io/projected/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-kube-api-access-cjfrl\") pod \"redhat-marketplace-9wgm4\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:07 crc kubenswrapper[4966]: I1217 10:49:07.124792 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:07 crc kubenswrapper[4966]: I1217 10:49:07.724083 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wgm4"] Dec 17 10:49:07 crc kubenswrapper[4966]: W1217 10:49:07.735806 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44c0fb2d_d35e_49ff_b57f_779c63a9ed40.slice/crio-58931bceca4c087e8f695b9f3eae0d9171a6af2d2d31ecf860d4ad469aea1269 WatchSource:0}: Error finding container 58931bceca4c087e8f695b9f3eae0d9171a6af2d2d31ecf860d4ad469aea1269: Status 404 returned error can't find the container with id 58931bceca4c087e8f695b9f3eae0d9171a6af2d2d31ecf860d4ad469aea1269 Dec 17 10:49:08 crc kubenswrapper[4966]: I1217 10:49:08.225565 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wgm4" event={"ID":"44c0fb2d-d35e-49ff-b57f-779c63a9ed40","Type":"ContainerDied","Data":"d991819ae51d33dc438fb70a496282bb354f4d47ae4732cb7bbce06dd8cb0f0c"} Dec 17 10:49:08 crc kubenswrapper[4966]: I1217 10:49:08.225477 4966 generic.go:334] "Generic (PLEG): container finished" podID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerID="d991819ae51d33dc438fb70a496282bb354f4d47ae4732cb7bbce06dd8cb0f0c" exitCode=0 Dec 17 10:49:08 crc kubenswrapper[4966]: I1217 10:49:08.227169 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wgm4" event={"ID":"44c0fb2d-d35e-49ff-b57f-779c63a9ed40","Type":"ContainerStarted","Data":"58931bceca4c087e8f695b9f3eae0d9171a6af2d2d31ecf860d4ad469aea1269"} Dec 17 10:49:08 crc kubenswrapper[4966]: I1217 10:49:08.228381 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 10:49:10 crc kubenswrapper[4966]: I1217 10:49:10.245094 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wgm4" event={"ID":"44c0fb2d-d35e-49ff-b57f-779c63a9ed40","Type":"ContainerStarted","Data":"023ea9440234c5524f6051c14a2c9b0182c0782a4e781fce508dcf1f938ccd27"} Dec 17 10:49:11 crc kubenswrapper[4966]: I1217 10:49:11.255014 4966 generic.go:334] "Generic (PLEG): container finished" podID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerID="023ea9440234c5524f6051c14a2c9b0182c0782a4e781fce508dcf1f938ccd27" exitCode=0 Dec 17 10:49:11 crc kubenswrapper[4966]: I1217 10:49:11.255121 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wgm4" event={"ID":"44c0fb2d-d35e-49ff-b57f-779c63a9ed40","Type":"ContainerDied","Data":"023ea9440234c5524f6051c14a2c9b0182c0782a4e781fce508dcf1f938ccd27"} Dec 17 10:49:12 crc kubenswrapper[4966]: I1217 10:49:12.267515 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wgm4" event={"ID":"44c0fb2d-d35e-49ff-b57f-779c63a9ed40","Type":"ContainerStarted","Data":"93380b3610aee03d742b3277d2e8b6ece5c420ce2f2caf046e81dffe9aa9e390"} Dec 17 10:49:12 crc kubenswrapper[4966]: I1217 10:49:12.293576 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9wgm4" podStartSLOduration=2.5566936780000002 podStartE2EDuration="6.293558103s" podCreationTimestamp="2025-12-17 10:49:06 +0000 UTC" firstStartedPulling="2025-12-17 10:49:08.226997573 +0000 UTC m=+8883.772067515" lastFinishedPulling="2025-12-17 10:49:11.963861998 +0000 UTC m=+8887.508931940" observedRunningTime="2025-12-17 10:49:12.290548181 +0000 UTC m=+8887.835618123" watchObservedRunningTime="2025-12-17 10:49:12.293558103 +0000 UTC m=+8887.838628045" Dec 17 10:49:16 crc kubenswrapper[4966]: I1217 10:49:16.807592 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:49:16 crc kubenswrapper[4966]: I1217 10:49:16.808183 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:49:16 crc kubenswrapper[4966]: I1217 10:49:16.808238 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:49:16 crc kubenswrapper[4966]: I1217 10:49:16.809138 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:49:16 crc kubenswrapper[4966]: I1217 10:49:16.809523 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" gracePeriod=600 Dec 17 10:49:16 crc kubenswrapper[4966]: E1217 10:49:16.931981 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:49:17 crc kubenswrapper[4966]: I1217 10:49:17.125309 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:17 crc kubenswrapper[4966]: I1217 10:49:17.125375 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:17 crc kubenswrapper[4966]: I1217 10:49:17.179943 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:17 crc kubenswrapper[4966]: I1217 10:49:17.314846 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" exitCode=0 Dec 17 10:49:17 crc kubenswrapper[4966]: I1217 10:49:17.315098 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5"} Dec 17 10:49:17 crc kubenswrapper[4966]: I1217 10:49:17.315317 4966 scope.go:117] "RemoveContainer" containerID="be0da4b76c123744b21b8ebc207645cfead54d433f111b12928f641bbcb96caf" Dec 17 10:49:17 crc kubenswrapper[4966]: I1217 10:49:17.315785 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:49:17 crc kubenswrapper[4966]: E1217 10:49:17.316322 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:49:17 crc kubenswrapper[4966]: I1217 10:49:17.399104 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.023763 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wgm4"] Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.025054 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9wgm4" podUID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerName="registry-server" containerID="cri-o://93380b3610aee03d742b3277d2e8b6ece5c420ce2f2caf046e81dffe9aa9e390" gracePeriod=2 Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.358995 4966 generic.go:334] "Generic (PLEG): container finished" podID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerID="93380b3610aee03d742b3277d2e8b6ece5c420ce2f2caf046e81dffe9aa9e390" exitCode=0 Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.359051 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wgm4" event={"ID":"44c0fb2d-d35e-49ff-b57f-779c63a9ed40","Type":"ContainerDied","Data":"93380b3610aee03d742b3277d2e8b6ece5c420ce2f2caf046e81dffe9aa9e390"} Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.646136 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.745634 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-utilities\") pod \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.745673 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-catalog-content\") pod \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.745797 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjfrl\" (UniqueName: \"kubernetes.io/projected/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-kube-api-access-cjfrl\") pod \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\" (UID: \"44c0fb2d-d35e-49ff-b57f-779c63a9ed40\") " Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.747121 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-utilities" (OuterVolumeSpecName: "utilities") pod "44c0fb2d-d35e-49ff-b57f-779c63a9ed40" (UID: "44c0fb2d-d35e-49ff-b57f-779c63a9ed40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.754280 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-kube-api-access-cjfrl" (OuterVolumeSpecName: "kube-api-access-cjfrl") pod "44c0fb2d-d35e-49ff-b57f-779c63a9ed40" (UID: "44c0fb2d-d35e-49ff-b57f-779c63a9ed40"). InnerVolumeSpecName "kube-api-access-cjfrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.766731 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44c0fb2d-d35e-49ff-b57f-779c63a9ed40" (UID: "44c0fb2d-d35e-49ff-b57f-779c63a9ed40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.847755 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.847819 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:49:21 crc kubenswrapper[4966]: I1217 10:49:21.847834 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjfrl\" (UniqueName: \"kubernetes.io/projected/44c0fb2d-d35e-49ff-b57f-779c63a9ed40-kube-api-access-cjfrl\") on node \"crc\" DevicePath \"\"" Dec 17 10:49:22 crc kubenswrapper[4966]: I1217 10:49:22.372065 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9wgm4" event={"ID":"44c0fb2d-d35e-49ff-b57f-779c63a9ed40","Type":"ContainerDied","Data":"58931bceca4c087e8f695b9f3eae0d9171a6af2d2d31ecf860d4ad469aea1269"} Dec 17 10:49:22 crc kubenswrapper[4966]: I1217 10:49:22.372129 4966 scope.go:117] "RemoveContainer" containerID="93380b3610aee03d742b3277d2e8b6ece5c420ce2f2caf046e81dffe9aa9e390" Dec 17 10:49:22 crc kubenswrapper[4966]: I1217 10:49:22.372291 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9wgm4" Dec 17 10:49:22 crc kubenswrapper[4966]: I1217 10:49:22.413445 4966 scope.go:117] "RemoveContainer" containerID="023ea9440234c5524f6051c14a2c9b0182c0782a4e781fce508dcf1f938ccd27" Dec 17 10:49:22 crc kubenswrapper[4966]: I1217 10:49:22.429568 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wgm4"] Dec 17 10:49:22 crc kubenswrapper[4966]: I1217 10:49:22.437773 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9wgm4"] Dec 17 10:49:22 crc kubenswrapper[4966]: I1217 10:49:22.438671 4966 scope.go:117] "RemoveContainer" containerID="d991819ae51d33dc438fb70a496282bb354f4d47ae4732cb7bbce06dd8cb0f0c" Dec 17 10:49:22 crc kubenswrapper[4966]: I1217 10:49:22.841335 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" path="/var/lib/kubelet/pods/44c0fb2d-d35e-49ff-b57f-779c63a9ed40/volumes" Dec 17 10:49:28 crc kubenswrapper[4966]: I1217 10:49:28.830983 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:49:28 crc kubenswrapper[4966]: E1217 10:49:28.831815 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:49:43 crc kubenswrapper[4966]: I1217 10:49:43.830138 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:49:43 crc kubenswrapper[4966]: E1217 10:49:43.831042 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.260207 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6f654f74c7-cpkpp"] Dec 17 10:49:54 crc kubenswrapper[4966]: E1217 10:49:54.261161 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerName="extract-utilities" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.261174 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerName="extract-utilities" Dec 17 10:49:54 crc kubenswrapper[4966]: E1217 10:49:54.261194 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerName="extract-content" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.261200 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerName="extract-content" Dec 17 10:49:54 crc kubenswrapper[4966]: E1217 10:49:54.261214 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerName="registry-server" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.261220 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerName="registry-server" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.261407 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c0fb2d-d35e-49ff-b57f-779c63a9ed40" containerName="registry-server" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.262441 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.287556 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f654f74c7-cpkpp"] Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.377626 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-config\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.378115 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-combined-ca-bundle\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.378171 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-public-tls-certs\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.378231 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-httpd-config\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.378267 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz7ws\" (UniqueName: \"kubernetes.io/projected/55525c8c-ab0d-4a6a-accf-9072864f5419-kube-api-access-wz7ws\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.378321 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-ovndb-tls-certs\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.378378 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-internal-tls-certs\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.486349 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-config\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.486457 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-combined-ca-bundle\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.486509 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-public-tls-certs\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.486572 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-httpd-config\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.486609 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz7ws\" (UniqueName: \"kubernetes.io/projected/55525c8c-ab0d-4a6a-accf-9072864f5419-kube-api-access-wz7ws\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.486659 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-ovndb-tls-certs\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.486710 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-internal-tls-certs\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.495146 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-internal-tls-certs\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.497220 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-combined-ca-bundle\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.501285 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-ovndb-tls-certs\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.505727 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-config\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.514277 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-httpd-config\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.515374 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525c8c-ab0d-4a6a-accf-9072864f5419-public-tls-certs\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.518501 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz7ws\" (UniqueName: \"kubernetes.io/projected/55525c8c-ab0d-4a6a-accf-9072864f5419-kube-api-access-wz7ws\") pod \"neutron-6f654f74c7-cpkpp\" (UID: \"55525c8c-ab0d-4a6a-accf-9072864f5419\") " pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:54 crc kubenswrapper[4966]: I1217 10:49:54.583373 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:55 crc kubenswrapper[4966]: I1217 10:49:55.436153 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f654f74c7-cpkpp"] Dec 17 10:49:55 crc kubenswrapper[4966]: I1217 10:49:55.665255 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f654f74c7-cpkpp" event={"ID":"55525c8c-ab0d-4a6a-accf-9072864f5419","Type":"ContainerStarted","Data":"458f5c1c7ad854e6d5d444aed2bc2566873f8cfd18fb405ef7e609c36daa2d02"} Dec 17 10:49:55 crc kubenswrapper[4966]: I1217 10:49:55.830765 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:49:55 crc kubenswrapper[4966]: E1217 10:49:55.831084 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:49:56 crc kubenswrapper[4966]: I1217 10:49:56.676967 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f654f74c7-cpkpp" event={"ID":"55525c8c-ab0d-4a6a-accf-9072864f5419","Type":"ContainerStarted","Data":"f2dce61a88a956e876e6a2a0bf1647953d6cdcd71f38ef7c1d23f5d1f6f6a324"} Dec 17 10:49:56 crc kubenswrapper[4966]: I1217 10:49:56.678268 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f654f74c7-cpkpp" event={"ID":"55525c8c-ab0d-4a6a-accf-9072864f5419","Type":"ContainerStarted","Data":"52c46cfbd799fad631840083a2194042334d520a48d28a8daa5c7b18ee8c8066"} Dec 17 10:49:56 crc kubenswrapper[4966]: I1217 10:49:56.678388 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:49:56 crc kubenswrapper[4966]: I1217 10:49:56.703956 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6f654f74c7-cpkpp" podStartSLOduration=2.703932761 podStartE2EDuration="2.703932761s" podCreationTimestamp="2025-12-17 10:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 10:49:56.696051737 +0000 UTC m=+8932.241121679" watchObservedRunningTime="2025-12-17 10:49:56.703932761 +0000 UTC m=+8932.249002713" Dec 17 10:50:09 crc kubenswrapper[4966]: I1217 10:50:09.830518 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:50:09 crc kubenswrapper[4966]: E1217 10:50:09.831331 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:50:21 crc kubenswrapper[4966]: I1217 10:50:21.831838 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:50:21 crc kubenswrapper[4966]: E1217 10:50:21.832558 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:50:24 crc kubenswrapper[4966]: I1217 10:50:24.599778 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6f654f74c7-cpkpp" Dec 17 10:50:24 crc kubenswrapper[4966]: I1217 10:50:24.689475 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-545d74fd97-8brzm"] Dec 17 10:50:24 crc kubenswrapper[4966]: I1217 10:50:24.689990 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-545d74fd97-8brzm" podUID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerName="neutron-api" containerID="cri-o://bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb" gracePeriod=30 Dec 17 10:50:24 crc kubenswrapper[4966]: I1217 10:50:24.690527 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-545d74fd97-8brzm" podUID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerName="neutron-httpd" containerID="cri-o://770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08" gracePeriod=30 Dec 17 10:50:25 crc kubenswrapper[4966]: I1217 10:50:25.936308 4966 generic.go:334] "Generic (PLEG): container finished" podID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerID="770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08" exitCode=0 Dec 17 10:50:25 crc kubenswrapper[4966]: I1217 10:50:25.936392 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-545d74fd97-8brzm" event={"ID":"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9","Type":"ContainerDied","Data":"770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08"} Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.738837 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.757300 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-whjx2"] Dec 17 10:50:29 crc kubenswrapper[4966]: E1217 10:50:29.757767 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerName="neutron-api" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.757792 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerName="neutron-api" Dec 17 10:50:29 crc kubenswrapper[4966]: E1217 10:50:29.757837 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerName="neutron-httpd" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.757844 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerName="neutron-httpd" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.758068 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerName="neutron-httpd" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.758090 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerName="neutron-api" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.759815 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.782896 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-whjx2"] Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.835241 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr6rm\" (UniqueName: \"kubernetes.io/projected/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-kube-api-access-qr6rm\") pod \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.835311 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-internal-tls-certs\") pod \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.835340 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-public-tls-certs\") pod \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.835529 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-config\") pod \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.835551 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-ovndb-tls-certs\") pod \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.835592 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-combined-ca-bundle\") pod \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.835672 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-httpd-config\") pod \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\" (UID: \"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9\") " Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.836494 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96m4n\" (UniqueName: \"kubernetes.io/projected/f664058f-43d5-470e-ba0b-8cdb68662a25-kube-api-access-96m4n\") pod \"certified-operators-whjx2\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.836534 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-catalog-content\") pod \"certified-operators-whjx2\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.836734 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-utilities\") pod \"certified-operators-whjx2\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.861672 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-kube-api-access-qr6rm" (OuterVolumeSpecName: "kube-api-access-qr6rm") pod "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" (UID: "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9"). InnerVolumeSpecName "kube-api-access-qr6rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.878511 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" (UID: "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.927368 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" (UID: "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.937593 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-config" (OuterVolumeSpecName: "config") pod "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" (UID: "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.942545 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-utilities\") pod \"certified-operators-whjx2\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.942664 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96m4n\" (UniqueName: \"kubernetes.io/projected/f664058f-43d5-470e-ba0b-8cdb68662a25-kube-api-access-96m4n\") pod \"certified-operators-whjx2\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.942706 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-catalog-content\") pod \"certified-operators-whjx2\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.942796 4966 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.942810 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr6rm\" (UniqueName: \"kubernetes.io/projected/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-kube-api-access-qr6rm\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.942823 4966 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.942834 4966 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-config\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.944327 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-utilities\") pod \"certified-operators-whjx2\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.945857 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-catalog-content\") pod \"certified-operators-whjx2\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.955787 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" (UID: "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.956053 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" (UID: "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.961187 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96m4n\" (UniqueName: \"kubernetes.io/projected/f664058f-43d5-470e-ba0b-8cdb68662a25-kube-api-access-96m4n\") pod \"certified-operators-whjx2\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.979749 4966 generic.go:334] "Generic (PLEG): container finished" podID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" containerID="bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb" exitCode=0 Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.979801 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-545d74fd97-8brzm" event={"ID":"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9","Type":"ContainerDied","Data":"bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb"} Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.979810 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-545d74fd97-8brzm" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.979830 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-545d74fd97-8brzm" event={"ID":"ac7ccc9b-4e29-41db-8d90-bcfba3271ef9","Type":"ContainerDied","Data":"4e267a69e5de66f2ee6287defc4bd12204333330f6627e0817988675e63fb79e"} Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.979851 4966 scope.go:117] "RemoveContainer" containerID="770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08" Dec 17 10:50:29 crc kubenswrapper[4966]: I1217 10:50:29.988105 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" (UID: "ac7ccc9b-4e29-41db-8d90-bcfba3271ef9"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.022015 4966 scope.go:117] "RemoveContainer" containerID="bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.043510 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.043550 4966 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.043561 4966 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.047190 4966 scope.go:117] "RemoveContainer" containerID="770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08" Dec 17 10:50:30 crc kubenswrapper[4966]: E1217 10:50:30.048546 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08\": container with ID starting with 770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08 not found: ID does not exist" containerID="770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.048713 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08"} err="failed to get container status \"770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08\": rpc error: code = NotFound desc = could not find container \"770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08\": container with ID starting with 770604007ba826e3139b233bc2b8bfc21db08bd7e35e8eb097c3a1d64bda2c08 not found: ID does not exist" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.048747 4966 scope.go:117] "RemoveContainer" containerID="bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb" Dec 17 10:50:30 crc kubenswrapper[4966]: E1217 10:50:30.049269 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb\": container with ID starting with bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb not found: ID does not exist" containerID="bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.049309 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb"} err="failed to get container status \"bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb\": rpc error: code = NotFound desc = could not find container \"bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb\": container with ID starting with bafe4bab82011f4710f3333c5e8662242aab577f58a0b28a18e9177c45eb90cb not found: ID does not exist" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.081031 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.339803 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-545d74fd97-8brzm"] Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.350857 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-545d74fd97-8brzm"] Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.616659 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-whjx2"] Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.843248 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac7ccc9b-4e29-41db-8d90-bcfba3271ef9" path="/var/lib/kubelet/pods/ac7ccc9b-4e29-41db-8d90-bcfba3271ef9/volumes" Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.997627 4966 generic.go:334] "Generic (PLEG): container finished" podID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerID="94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d" exitCode=0 Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.997742 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-whjx2" event={"ID":"f664058f-43d5-470e-ba0b-8cdb68662a25","Type":"ContainerDied","Data":"94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d"} Dec 17 10:50:30 crc kubenswrapper[4966]: I1217 10:50:30.997782 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-whjx2" event={"ID":"f664058f-43d5-470e-ba0b-8cdb68662a25","Type":"ContainerStarted","Data":"aaa35da88e195b9d636244374508c4104ae7cc0a746bb32672c0f4ce746569ba"} Dec 17 10:50:32 crc kubenswrapper[4966]: I1217 10:50:32.833791 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:50:32 crc kubenswrapper[4966]: E1217 10:50:32.835480 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:50:33 crc kubenswrapper[4966]: I1217 10:50:33.019276 4966 generic.go:334] "Generic (PLEG): container finished" podID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerID="c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939" exitCode=0 Dec 17 10:50:33 crc kubenswrapper[4966]: I1217 10:50:33.019336 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-whjx2" event={"ID":"f664058f-43d5-470e-ba0b-8cdb68662a25","Type":"ContainerDied","Data":"c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939"} Dec 17 10:50:34 crc kubenswrapper[4966]: I1217 10:50:34.030222 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-whjx2" event={"ID":"f664058f-43d5-470e-ba0b-8cdb68662a25","Type":"ContainerStarted","Data":"9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5"} Dec 17 10:50:34 crc kubenswrapper[4966]: I1217 10:50:34.049545 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-whjx2" podStartSLOduration=2.596282391 podStartE2EDuration="5.049528236s" podCreationTimestamp="2025-12-17 10:50:29 +0000 UTC" firstStartedPulling="2025-12-17 10:50:31.000781663 +0000 UTC m=+8966.545851615" lastFinishedPulling="2025-12-17 10:50:33.454027518 +0000 UTC m=+8968.999097460" observedRunningTime="2025-12-17 10:50:34.046171344 +0000 UTC m=+8969.591241306" watchObservedRunningTime="2025-12-17 10:50:34.049528236 +0000 UTC m=+8969.594598178" Dec 17 10:50:40 crc kubenswrapper[4966]: I1217 10:50:40.081900 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:40 crc kubenswrapper[4966]: I1217 10:50:40.083279 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:40 crc kubenswrapper[4966]: I1217 10:50:40.126998 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:41 crc kubenswrapper[4966]: I1217 10:50:41.138929 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:41 crc kubenswrapper[4966]: I1217 10:50:41.196618 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-whjx2"] Dec 17 10:50:43 crc kubenswrapper[4966]: I1217 10:50:43.111022 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-whjx2" podUID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerName="registry-server" containerID="cri-o://9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5" gracePeriod=2 Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.068291 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.122127 4966 generic.go:334] "Generic (PLEG): container finished" podID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerID="9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5" exitCode=0 Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.122172 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-whjx2" event={"ID":"f664058f-43d5-470e-ba0b-8cdb68662a25","Type":"ContainerDied","Data":"9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5"} Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.122190 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-whjx2" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.122210 4966 scope.go:117] "RemoveContainer" containerID="9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.122198 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-whjx2" event={"ID":"f664058f-43d5-470e-ba0b-8cdb68662a25","Type":"ContainerDied","Data":"aaa35da88e195b9d636244374508c4104ae7cc0a746bb32672c0f4ce746569ba"} Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.136785 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-catalog-content\") pod \"f664058f-43d5-470e-ba0b-8cdb68662a25\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.137214 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96m4n\" (UniqueName: \"kubernetes.io/projected/f664058f-43d5-470e-ba0b-8cdb68662a25-kube-api-access-96m4n\") pod \"f664058f-43d5-470e-ba0b-8cdb68662a25\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.137401 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-utilities\") pod \"f664058f-43d5-470e-ba0b-8cdb68662a25\" (UID: \"f664058f-43d5-470e-ba0b-8cdb68662a25\") " Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.138740 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-utilities" (OuterVolumeSpecName: "utilities") pod "f664058f-43d5-470e-ba0b-8cdb68662a25" (UID: "f664058f-43d5-470e-ba0b-8cdb68662a25"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.154698 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f664058f-43d5-470e-ba0b-8cdb68662a25-kube-api-access-96m4n" (OuterVolumeSpecName: "kube-api-access-96m4n") pod "f664058f-43d5-470e-ba0b-8cdb68662a25" (UID: "f664058f-43d5-470e-ba0b-8cdb68662a25"). InnerVolumeSpecName "kube-api-access-96m4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.162123 4966 scope.go:117] "RemoveContainer" containerID="c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.213223 4966 scope.go:117] "RemoveContainer" containerID="94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.213569 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f664058f-43d5-470e-ba0b-8cdb68662a25" (UID: "f664058f-43d5-470e-ba0b-8cdb68662a25"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.239373 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.239407 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96m4n\" (UniqueName: \"kubernetes.io/projected/f664058f-43d5-470e-ba0b-8cdb68662a25-kube-api-access-96m4n\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.239418 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f664058f-43d5-470e-ba0b-8cdb68662a25-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.260462 4966 scope.go:117] "RemoveContainer" containerID="9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5" Dec 17 10:50:44 crc kubenswrapper[4966]: E1217 10:50:44.261985 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5\": container with ID starting with 9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5 not found: ID does not exist" containerID="9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.262020 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5"} err="failed to get container status \"9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5\": rpc error: code = NotFound desc = could not find container \"9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5\": container with ID starting with 9e5148c5ab293788ccf1652e60085537362cdf2251bf9b5bd1993185d5d7d8f5 not found: ID does not exist" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.262041 4966 scope.go:117] "RemoveContainer" containerID="c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939" Dec 17 10:50:44 crc kubenswrapper[4966]: E1217 10:50:44.262635 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939\": container with ID starting with c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939 not found: ID does not exist" containerID="c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.262683 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939"} err="failed to get container status \"c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939\": rpc error: code = NotFound desc = could not find container \"c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939\": container with ID starting with c4b764fad87ed746f2a2734802978a2dab3d591701bd72ea4a7db476b2c66939 not found: ID does not exist" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.262718 4966 scope.go:117] "RemoveContainer" containerID="94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d" Dec 17 10:50:44 crc kubenswrapper[4966]: E1217 10:50:44.263144 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d\": container with ID starting with 94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d not found: ID does not exist" containerID="94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.263170 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d"} err="failed to get container status \"94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d\": rpc error: code = NotFound desc = could not find container \"94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d\": container with ID starting with 94db3cf41e0baac9f0273d7833bd717c610a11f0455e9d6f15d2ee83bbf7fa9d not found: ID does not exist" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.455234 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-whjx2"] Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.474304 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-whjx2"] Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.836904 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:50:44 crc kubenswrapper[4966]: E1217 10:50:44.837315 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:50:44 crc kubenswrapper[4966]: I1217 10:50:44.841862 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f664058f-43d5-470e-ba0b-8cdb68662a25" path="/var/lib/kubelet/pods/f664058f-43d5-470e-ba0b-8cdb68662a25/volumes" Dec 17 10:50:57 crc kubenswrapper[4966]: I1217 10:50:57.830395 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:50:57 crc kubenswrapper[4966]: E1217 10:50:57.831188 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:51:08 crc kubenswrapper[4966]: I1217 10:51:08.951130 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8crhr"] Dec 17 10:51:08 crc kubenswrapper[4966]: E1217 10:51:08.952065 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerName="registry-server" Dec 17 10:51:08 crc kubenswrapper[4966]: I1217 10:51:08.952077 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerName="registry-server" Dec 17 10:51:08 crc kubenswrapper[4966]: E1217 10:51:08.952110 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerName="extract-utilities" Dec 17 10:51:08 crc kubenswrapper[4966]: I1217 10:51:08.952117 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerName="extract-utilities" Dec 17 10:51:08 crc kubenswrapper[4966]: E1217 10:51:08.952135 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerName="extract-content" Dec 17 10:51:08 crc kubenswrapper[4966]: I1217 10:51:08.952142 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerName="extract-content" Dec 17 10:51:08 crc kubenswrapper[4966]: I1217 10:51:08.952383 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="f664058f-43d5-470e-ba0b-8cdb68662a25" containerName="registry-server" Dec 17 10:51:08 crc kubenswrapper[4966]: I1217 10:51:08.954246 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:08 crc kubenswrapper[4966]: I1217 10:51:08.972307 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8crhr"] Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.140465 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-catalog-content\") pod \"community-operators-8crhr\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.140534 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jqnw\" (UniqueName: \"kubernetes.io/projected/c0100497-0a66-4967-a7cb-af8215d663c3-kube-api-access-8jqnw\") pod \"community-operators-8crhr\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.140693 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-utilities\") pod \"community-operators-8crhr\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.242573 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-catalog-content\") pod \"community-operators-8crhr\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.242632 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jqnw\" (UniqueName: \"kubernetes.io/projected/c0100497-0a66-4967-a7cb-af8215d663c3-kube-api-access-8jqnw\") pod \"community-operators-8crhr\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.242698 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-utilities\") pod \"community-operators-8crhr\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.243138 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-catalog-content\") pod \"community-operators-8crhr\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.243147 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-utilities\") pod \"community-operators-8crhr\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.268352 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jqnw\" (UniqueName: \"kubernetes.io/projected/c0100497-0a66-4967-a7cb-af8215d663c3-kube-api-access-8jqnw\") pod \"community-operators-8crhr\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.274817 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:09 crc kubenswrapper[4966]: I1217 10:51:09.965469 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8crhr"] Dec 17 10:51:10 crc kubenswrapper[4966]: I1217 10:51:10.365940 4966 generic.go:334] "Generic (PLEG): container finished" podID="c0100497-0a66-4967-a7cb-af8215d663c3" containerID="36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a" exitCode=0 Dec 17 10:51:10 crc kubenswrapper[4966]: I1217 10:51:10.366049 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8crhr" event={"ID":"c0100497-0a66-4967-a7cb-af8215d663c3","Type":"ContainerDied","Data":"36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a"} Dec 17 10:51:10 crc kubenswrapper[4966]: I1217 10:51:10.366336 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8crhr" event={"ID":"c0100497-0a66-4967-a7cb-af8215d663c3","Type":"ContainerStarted","Data":"adc841c3fce6bfb0a5987d7dcbebad89121c9c69d369571cbc9d1f2632831a3c"} Dec 17 10:51:10 crc kubenswrapper[4966]: I1217 10:51:10.830947 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:51:10 crc kubenswrapper[4966]: E1217 10:51:10.831201 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:51:12 crc kubenswrapper[4966]: I1217 10:51:12.395990 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8crhr" event={"ID":"c0100497-0a66-4967-a7cb-af8215d663c3","Type":"ContainerStarted","Data":"ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b"} Dec 17 10:51:13 crc kubenswrapper[4966]: I1217 10:51:13.406356 4966 generic.go:334] "Generic (PLEG): container finished" podID="c0100497-0a66-4967-a7cb-af8215d663c3" containerID="ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b" exitCode=0 Dec 17 10:51:13 crc kubenswrapper[4966]: I1217 10:51:13.406414 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8crhr" event={"ID":"c0100497-0a66-4967-a7cb-af8215d663c3","Type":"ContainerDied","Data":"ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b"} Dec 17 10:51:14 crc kubenswrapper[4966]: I1217 10:51:14.429677 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8crhr" event={"ID":"c0100497-0a66-4967-a7cb-af8215d663c3","Type":"ContainerStarted","Data":"ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd"} Dec 17 10:51:14 crc kubenswrapper[4966]: I1217 10:51:14.461441 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8crhr" podStartSLOduration=3.016364218 podStartE2EDuration="6.46141542s" podCreationTimestamp="2025-12-17 10:51:08 +0000 UTC" firstStartedPulling="2025-12-17 10:51:10.367692051 +0000 UTC m=+9005.912761993" lastFinishedPulling="2025-12-17 10:51:13.812743253 +0000 UTC m=+9009.357813195" observedRunningTime="2025-12-17 10:51:14.453428053 +0000 UTC m=+9009.998497995" watchObservedRunningTime="2025-12-17 10:51:14.46141542 +0000 UTC m=+9010.006485362" Dec 17 10:51:19 crc kubenswrapper[4966]: I1217 10:51:19.275982 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:19 crc kubenswrapper[4966]: I1217 10:51:19.276513 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:19 crc kubenswrapper[4966]: I1217 10:51:19.362974 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:19 crc kubenswrapper[4966]: I1217 10:51:19.519935 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:19 crc kubenswrapper[4966]: I1217 10:51:19.605276 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8crhr"] Dec 17 10:51:21 crc kubenswrapper[4966]: I1217 10:51:21.492953 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8crhr" podUID="c0100497-0a66-4967-a7cb-af8215d663c3" containerName="registry-server" containerID="cri-o://ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd" gracePeriod=2 Dec 17 10:51:21 crc kubenswrapper[4966]: I1217 10:51:21.830209 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:51:21 crc kubenswrapper[4966]: E1217 10:51:21.830854 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.044697 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.194927 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jqnw\" (UniqueName: \"kubernetes.io/projected/c0100497-0a66-4967-a7cb-af8215d663c3-kube-api-access-8jqnw\") pod \"c0100497-0a66-4967-a7cb-af8215d663c3\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.195448 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-catalog-content\") pod \"c0100497-0a66-4967-a7cb-af8215d663c3\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.195638 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-utilities\") pod \"c0100497-0a66-4967-a7cb-af8215d663c3\" (UID: \"c0100497-0a66-4967-a7cb-af8215d663c3\") " Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.196439 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-utilities" (OuterVolumeSpecName: "utilities") pod "c0100497-0a66-4967-a7cb-af8215d663c3" (UID: "c0100497-0a66-4967-a7cb-af8215d663c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.203279 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0100497-0a66-4967-a7cb-af8215d663c3-kube-api-access-8jqnw" (OuterVolumeSpecName: "kube-api-access-8jqnw") pod "c0100497-0a66-4967-a7cb-af8215d663c3" (UID: "c0100497-0a66-4967-a7cb-af8215d663c3"). InnerVolumeSpecName "kube-api-access-8jqnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.269501 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0100497-0a66-4967-a7cb-af8215d663c3" (UID: "c0100497-0a66-4967-a7cb-af8215d663c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.298568 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.298615 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jqnw\" (UniqueName: \"kubernetes.io/projected/c0100497-0a66-4967-a7cb-af8215d663c3-kube-api-access-8jqnw\") on node \"crc\" DevicePath \"\"" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.298633 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0100497-0a66-4967-a7cb-af8215d663c3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.502478 4966 generic.go:334] "Generic (PLEG): container finished" podID="c0100497-0a66-4967-a7cb-af8215d663c3" containerID="ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd" exitCode=0 Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.502715 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8crhr" event={"ID":"c0100497-0a66-4967-a7cb-af8215d663c3","Type":"ContainerDied","Data":"ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd"} Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.502760 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8crhr" event={"ID":"c0100497-0a66-4967-a7cb-af8215d663c3","Type":"ContainerDied","Data":"adc841c3fce6bfb0a5987d7dcbebad89121c9c69d369571cbc9d1f2632831a3c"} Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.502766 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8crhr" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.502784 4966 scope.go:117] "RemoveContainer" containerID="ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.538167 4966 scope.go:117] "RemoveContainer" containerID="ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.569021 4966 scope.go:117] "RemoveContainer" containerID="36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.578139 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8crhr"] Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.588265 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8crhr"] Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.623141 4966 scope.go:117] "RemoveContainer" containerID="ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd" Dec 17 10:51:22 crc kubenswrapper[4966]: E1217 10:51:22.623951 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd\": container with ID starting with ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd not found: ID does not exist" containerID="ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.623982 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd"} err="failed to get container status \"ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd\": rpc error: code = NotFound desc = could not find container \"ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd\": container with ID starting with ee6f1b1575bed7eb8471b67dcd960992fc13f23bdc29a2d62a3d9c550b6472cd not found: ID does not exist" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.624004 4966 scope.go:117] "RemoveContainer" containerID="ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b" Dec 17 10:51:22 crc kubenswrapper[4966]: E1217 10:51:22.624743 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b\": container with ID starting with ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b not found: ID does not exist" containerID="ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.624795 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b"} err="failed to get container status \"ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b\": rpc error: code = NotFound desc = could not find container \"ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b\": container with ID starting with ff11aef62cd1fc68b4d23117e390a6733ae235b39ff71654bae519e05225193b not found: ID does not exist" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.624829 4966 scope.go:117] "RemoveContainer" containerID="36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a" Dec 17 10:51:22 crc kubenswrapper[4966]: E1217 10:51:22.625167 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a\": container with ID starting with 36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a not found: ID does not exist" containerID="36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.625203 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a"} err="failed to get container status \"36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a\": rpc error: code = NotFound desc = could not find container \"36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a\": container with ID starting with 36dc7a60e6d741595aea7613931bdc237d040e401a3e2aaeb331fbac6c95208a not found: ID does not exist" Dec 17 10:51:22 crc kubenswrapper[4966]: I1217 10:51:22.842391 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0100497-0a66-4967-a7cb-af8215d663c3" path="/var/lib/kubelet/pods/c0100497-0a66-4967-a7cb-af8215d663c3/volumes" Dec 17 10:51:33 crc kubenswrapper[4966]: I1217 10:51:33.830782 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:51:33 crc kubenswrapper[4966]: E1217 10:51:33.832202 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:51:45 crc kubenswrapper[4966]: I1217 10:51:45.830386 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:51:45 crc kubenswrapper[4966]: E1217 10:51:45.831019 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:51:59 crc kubenswrapper[4966]: I1217 10:51:59.830073 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:51:59 crc kubenswrapper[4966]: E1217 10:51:59.830809 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:52:13 crc kubenswrapper[4966]: I1217 10:52:13.053529 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:52:13 crc kubenswrapper[4966]: E1217 10:52:13.054302 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:52:24 crc kubenswrapper[4966]: I1217 10:52:24.838093 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:52:24 crc kubenswrapper[4966]: E1217 10:52:24.839078 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:52:37 crc kubenswrapper[4966]: I1217 10:52:37.831773 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:52:37 crc kubenswrapper[4966]: E1217 10:52:37.832907 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:52:50 crc kubenswrapper[4966]: I1217 10:52:50.831204 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:52:50 crc kubenswrapper[4966]: E1217 10:52:50.832016 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:53:02 crc kubenswrapper[4966]: I1217 10:53:02.830920 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:53:02 crc kubenswrapper[4966]: E1217 10:53:02.832788 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:53:14 crc kubenswrapper[4966]: I1217 10:53:14.850945 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:53:14 crc kubenswrapper[4966]: E1217 10:53:14.852380 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:53:25 crc kubenswrapper[4966]: I1217 10:53:25.830976 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:53:25 crc kubenswrapper[4966]: E1217 10:53:25.832590 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:53:39 crc kubenswrapper[4966]: I1217 10:53:39.831104 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:53:39 crc kubenswrapper[4966]: E1217 10:53:39.832072 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:53:52 crc kubenswrapper[4966]: I1217 10:53:52.831018 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:53:52 crc kubenswrapper[4966]: E1217 10:53:52.831846 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:54:04 crc kubenswrapper[4966]: I1217 10:54:04.840208 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:54:04 crc kubenswrapper[4966]: E1217 10:54:04.841135 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:54:15 crc kubenswrapper[4966]: I1217 10:54:15.830474 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:54:15 crc kubenswrapper[4966]: E1217 10:54:15.832961 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 10:54:27 crc kubenswrapper[4966]: I1217 10:54:27.831068 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 10:54:28 crc kubenswrapper[4966]: I1217 10:54:28.317305 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"05f4c1c9032aed889945ebfbf97186e7dfeda826202a490e0f1d86c1a437ed61"} Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.892612 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jx52z"] Dec 17 10:55:58 crc kubenswrapper[4966]: E1217 10:55:58.893578 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0100497-0a66-4967-a7cb-af8215d663c3" containerName="registry-server" Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.893591 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0100497-0a66-4967-a7cb-af8215d663c3" containerName="registry-server" Dec 17 10:55:58 crc kubenswrapper[4966]: E1217 10:55:58.893614 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0100497-0a66-4967-a7cb-af8215d663c3" containerName="extract-content" Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.893620 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0100497-0a66-4967-a7cb-af8215d663c3" containerName="extract-content" Dec 17 10:55:58 crc kubenswrapper[4966]: E1217 10:55:58.893637 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0100497-0a66-4967-a7cb-af8215d663c3" containerName="extract-utilities" Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.893643 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0100497-0a66-4967-a7cb-af8215d663c3" containerName="extract-utilities" Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.893851 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0100497-0a66-4967-a7cb-af8215d663c3" containerName="registry-server" Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.895204 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.916375 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jx52z"] Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.991544 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxbpq\" (UniqueName: \"kubernetes.io/projected/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-kube-api-access-dxbpq\") pod \"redhat-operators-jx52z\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.992086 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-catalog-content\") pod \"redhat-operators-jx52z\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:58 crc kubenswrapper[4966]: I1217 10:55:58.992233 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-utilities\") pod \"redhat-operators-jx52z\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:59 crc kubenswrapper[4966]: I1217 10:55:59.093953 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-catalog-content\") pod \"redhat-operators-jx52z\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:59 crc kubenswrapper[4966]: I1217 10:55:59.094371 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-utilities\") pod \"redhat-operators-jx52z\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:59 crc kubenswrapper[4966]: I1217 10:55:59.094606 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxbpq\" (UniqueName: \"kubernetes.io/projected/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-kube-api-access-dxbpq\") pod \"redhat-operators-jx52z\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:59 crc kubenswrapper[4966]: I1217 10:55:59.094598 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-catalog-content\") pod \"redhat-operators-jx52z\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:59 crc kubenswrapper[4966]: I1217 10:55:59.094815 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-utilities\") pod \"redhat-operators-jx52z\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:59 crc kubenswrapper[4966]: I1217 10:55:59.124839 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxbpq\" (UniqueName: \"kubernetes.io/projected/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-kube-api-access-dxbpq\") pod \"redhat-operators-jx52z\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:59 crc kubenswrapper[4966]: I1217 10:55:59.222982 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:55:59 crc kubenswrapper[4966]: I1217 10:55:59.805487 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jx52z"] Dec 17 10:56:00 crc kubenswrapper[4966]: I1217 10:56:00.227894 4966 generic.go:334] "Generic (PLEG): container finished" podID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerID="7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d" exitCode=0 Dec 17 10:56:00 crc kubenswrapper[4966]: I1217 10:56:00.227938 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jx52z" event={"ID":"a1c65e73-91c4-437c-abb3-7cd4eb62a56f","Type":"ContainerDied","Data":"7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d"} Dec 17 10:56:00 crc kubenswrapper[4966]: I1217 10:56:00.227981 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jx52z" event={"ID":"a1c65e73-91c4-437c-abb3-7cd4eb62a56f","Type":"ContainerStarted","Data":"d061c156cf174673c344fc86564559504e3b6c4c8472309f092bdcd9d5d98fe9"} Dec 17 10:56:00 crc kubenswrapper[4966]: I1217 10:56:00.229929 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 10:56:02 crc kubenswrapper[4966]: I1217 10:56:02.249056 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jx52z" event={"ID":"a1c65e73-91c4-437c-abb3-7cd4eb62a56f","Type":"ContainerStarted","Data":"30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd"} Dec 17 10:56:05 crc kubenswrapper[4966]: I1217 10:56:05.273724 4966 generic.go:334] "Generic (PLEG): container finished" podID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerID="30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd" exitCode=0 Dec 17 10:56:05 crc kubenswrapper[4966]: I1217 10:56:05.273797 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jx52z" event={"ID":"a1c65e73-91c4-437c-abb3-7cd4eb62a56f","Type":"ContainerDied","Data":"30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd"} Dec 17 10:56:06 crc kubenswrapper[4966]: I1217 10:56:06.286045 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jx52z" event={"ID":"a1c65e73-91c4-437c-abb3-7cd4eb62a56f","Type":"ContainerStarted","Data":"c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073"} Dec 17 10:56:06 crc kubenswrapper[4966]: I1217 10:56:06.314418 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jx52z" podStartSLOduration=2.567220427 podStartE2EDuration="8.314395455s" podCreationTimestamp="2025-12-17 10:55:58 +0000 UTC" firstStartedPulling="2025-12-17 10:56:00.229483033 +0000 UTC m=+9295.774552975" lastFinishedPulling="2025-12-17 10:56:05.976658061 +0000 UTC m=+9301.521728003" observedRunningTime="2025-12-17 10:56:06.305975625 +0000 UTC m=+9301.851045567" watchObservedRunningTime="2025-12-17 10:56:06.314395455 +0000 UTC m=+9301.859465417" Dec 17 10:56:09 crc kubenswrapper[4966]: I1217 10:56:09.223627 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:56:09 crc kubenswrapper[4966]: I1217 10:56:09.224236 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:56:10 crc kubenswrapper[4966]: I1217 10:56:10.271247 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jx52z" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="registry-server" probeResult="failure" output=< Dec 17 10:56:10 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:56:10 crc kubenswrapper[4966]: > Dec 17 10:56:20 crc kubenswrapper[4966]: I1217 10:56:20.288996 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jx52z" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="registry-server" probeResult="failure" output=< Dec 17 10:56:20 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 10:56:20 crc kubenswrapper[4966]: > Dec 17 10:56:29 crc kubenswrapper[4966]: I1217 10:56:29.303638 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:56:29 crc kubenswrapper[4966]: I1217 10:56:29.367413 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:56:29 crc kubenswrapper[4966]: I1217 10:56:29.547982 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jx52z"] Dec 17 10:56:30 crc kubenswrapper[4966]: I1217 10:56:30.499521 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jx52z" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="registry-server" containerID="cri-o://c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073" gracePeriod=2 Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.150741 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.273974 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxbpq\" (UniqueName: \"kubernetes.io/projected/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-kube-api-access-dxbpq\") pod \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.274601 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-utilities\") pod \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.274838 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-catalog-content\") pod \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\" (UID: \"a1c65e73-91c4-437c-abb3-7cd4eb62a56f\") " Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.275195 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-utilities" (OuterVolumeSpecName: "utilities") pod "a1c65e73-91c4-437c-abb3-7cd4eb62a56f" (UID: "a1c65e73-91c4-437c-abb3-7cd4eb62a56f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.275827 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.281584 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-kube-api-access-dxbpq" (OuterVolumeSpecName: "kube-api-access-dxbpq") pod "a1c65e73-91c4-437c-abb3-7cd4eb62a56f" (UID: "a1c65e73-91c4-437c-abb3-7cd4eb62a56f"). InnerVolumeSpecName "kube-api-access-dxbpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.378540 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxbpq\" (UniqueName: \"kubernetes.io/projected/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-kube-api-access-dxbpq\") on node \"crc\" DevicePath \"\"" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.380925 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1c65e73-91c4-437c-abb3-7cd4eb62a56f" (UID: "a1c65e73-91c4-437c-abb3-7cd4eb62a56f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.480312 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c65e73-91c4-437c-abb3-7cd4eb62a56f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.509609 4966 generic.go:334] "Generic (PLEG): container finished" podID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerID="c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073" exitCode=0 Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.509660 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jx52z" event={"ID":"a1c65e73-91c4-437c-abb3-7cd4eb62a56f","Type":"ContainerDied","Data":"c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073"} Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.509701 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jx52z" event={"ID":"a1c65e73-91c4-437c-abb3-7cd4eb62a56f","Type":"ContainerDied","Data":"d061c156cf174673c344fc86564559504e3b6c4c8472309f092bdcd9d5d98fe9"} Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.509719 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jx52z" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.510022 4966 scope.go:117] "RemoveContainer" containerID="c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.563697 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jx52z"] Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.566690 4966 scope.go:117] "RemoveContainer" containerID="30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.572382 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jx52z"] Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.597494 4966 scope.go:117] "RemoveContainer" containerID="7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.643544 4966 scope.go:117] "RemoveContainer" containerID="c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073" Dec 17 10:56:31 crc kubenswrapper[4966]: E1217 10:56:31.644640 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073\": container with ID starting with c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073 not found: ID does not exist" containerID="c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.644686 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073"} err="failed to get container status \"c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073\": rpc error: code = NotFound desc = could not find container \"c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073\": container with ID starting with c7406a6090138fa114e749ff9172c708b386cdad3bb684cad3f4d0a668be7073 not found: ID does not exist" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.644713 4966 scope.go:117] "RemoveContainer" containerID="30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd" Dec 17 10:56:31 crc kubenswrapper[4966]: E1217 10:56:31.645156 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd\": container with ID starting with 30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd not found: ID does not exist" containerID="30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.645215 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd"} err="failed to get container status \"30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd\": rpc error: code = NotFound desc = could not find container \"30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd\": container with ID starting with 30d05172ee8a43400b80ce857d9c92e9af4e7b1c786d4380a519509b9ea59ecd not found: ID does not exist" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.645244 4966 scope.go:117] "RemoveContainer" containerID="7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d" Dec 17 10:56:31 crc kubenswrapper[4966]: E1217 10:56:31.647098 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d\": container with ID starting with 7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d not found: ID does not exist" containerID="7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d" Dec 17 10:56:31 crc kubenswrapper[4966]: I1217 10:56:31.647155 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d"} err="failed to get container status \"7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d\": rpc error: code = NotFound desc = could not find container \"7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d\": container with ID starting with 7a641717646c8995e781f127fd7634215b88757507a817ef781f60efeab9514d not found: ID does not exist" Dec 17 10:56:31 crc kubenswrapper[4966]: E1217 10:56:31.748361 4966 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1c65e73_91c4_437c_abb3_7cd4eb62a56f.slice/crio-d061c156cf174673c344fc86564559504e3b6c4c8472309f092bdcd9d5d98fe9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1c65e73_91c4_437c_abb3_7cd4eb62a56f.slice\": RecentStats: unable to find data in memory cache]" Dec 17 10:56:32 crc kubenswrapper[4966]: I1217 10:56:32.842991 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" path="/var/lib/kubelet/pods/a1c65e73-91c4-437c-abb3-7cd4eb62a56f/volumes" Dec 17 10:56:46 crc kubenswrapper[4966]: I1217 10:56:46.808177 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:56:46 crc kubenswrapper[4966]: I1217 10:56:46.809111 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:57:16 crc kubenswrapper[4966]: I1217 10:57:16.807997 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:57:16 crc kubenswrapper[4966]: I1217 10:57:16.808476 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:57:46 crc kubenswrapper[4966]: I1217 10:57:46.808149 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 10:57:46 crc kubenswrapper[4966]: I1217 10:57:46.808759 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 10:57:46 crc kubenswrapper[4966]: I1217 10:57:46.808854 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 10:57:46 crc kubenswrapper[4966]: I1217 10:57:46.809721 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"05f4c1c9032aed889945ebfbf97186e7dfeda826202a490e0f1d86c1a437ed61"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 10:57:46 crc kubenswrapper[4966]: I1217 10:57:46.809774 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://05f4c1c9032aed889945ebfbf97186e7dfeda826202a490e0f1d86c1a437ed61" gracePeriod=600 Dec 17 10:57:47 crc kubenswrapper[4966]: I1217 10:57:47.260843 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="05f4c1c9032aed889945ebfbf97186e7dfeda826202a490e0f1d86c1a437ed61" exitCode=0 Dec 17 10:57:47 crc kubenswrapper[4966]: I1217 10:57:47.261215 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"05f4c1c9032aed889945ebfbf97186e7dfeda826202a490e0f1d86c1a437ed61"} Dec 17 10:57:47 crc kubenswrapper[4966]: I1217 10:57:47.261244 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b"} Dec 17 10:57:47 crc kubenswrapper[4966]: I1217 10:57:47.261263 4966 scope.go:117] "RemoveContainer" containerID="571ccd90a29ae866b434bb4407448f429b47750575dab2596172d212696bc3b5" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.180119 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c"] Dec 17 11:00:00 crc kubenswrapper[4966]: E1217 11:00:00.180934 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="extract-utilities" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.180946 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="extract-utilities" Dec 17 11:00:00 crc kubenswrapper[4966]: E1217 11:00:00.180961 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="registry-server" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.180967 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="registry-server" Dec 17 11:00:00 crc kubenswrapper[4966]: E1217 11:00:00.180989 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="extract-content" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.180994 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="extract-content" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.181210 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c65e73-91c4-437c-abb3-7cd4eb62a56f" containerName="registry-server" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.182406 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.190844 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c"] Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.196281 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.266827 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.360518 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e5451e1-267f-4401-8d38-e02d8a6f8191-config-volume\") pod \"collect-profiles-29432820-wxk7c\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.360649 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2s6d\" (UniqueName: \"kubernetes.io/projected/8e5451e1-267f-4401-8d38-e02d8a6f8191-kube-api-access-t2s6d\") pod \"collect-profiles-29432820-wxk7c\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.360682 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e5451e1-267f-4401-8d38-e02d8a6f8191-secret-volume\") pod \"collect-profiles-29432820-wxk7c\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.462094 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e5451e1-267f-4401-8d38-e02d8a6f8191-config-volume\") pod \"collect-profiles-29432820-wxk7c\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.462225 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2s6d\" (UniqueName: \"kubernetes.io/projected/8e5451e1-267f-4401-8d38-e02d8a6f8191-kube-api-access-t2s6d\") pod \"collect-profiles-29432820-wxk7c\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.462254 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e5451e1-267f-4401-8d38-e02d8a6f8191-secret-volume\") pod \"collect-profiles-29432820-wxk7c\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.463866 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e5451e1-267f-4401-8d38-e02d8a6f8191-config-volume\") pod \"collect-profiles-29432820-wxk7c\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.498042 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e5451e1-267f-4401-8d38-e02d8a6f8191-secret-volume\") pod \"collect-profiles-29432820-wxk7c\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.498583 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2s6d\" (UniqueName: \"kubernetes.io/projected/8e5451e1-267f-4401-8d38-e02d8a6f8191-kube-api-access-t2s6d\") pod \"collect-profiles-29432820-wxk7c\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.509189 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.910097 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bgkll"] Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.912689 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:00 crc kubenswrapper[4966]: I1217 11:00:00.932436 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgkll"] Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:00.999996 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c"] Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.073977 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-catalog-content\") pod \"redhat-marketplace-bgkll\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.074541 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-utilities\") pod \"redhat-marketplace-bgkll\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.074725 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gptm8\" (UniqueName: \"kubernetes.io/projected/3870ac16-6934-481b-b91f-dea07612efe4-kube-api-access-gptm8\") pod \"redhat-marketplace-bgkll\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.176828 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-catalog-content\") pod \"redhat-marketplace-bgkll\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.176921 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-utilities\") pod \"redhat-marketplace-bgkll\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.176977 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gptm8\" (UniqueName: \"kubernetes.io/projected/3870ac16-6934-481b-b91f-dea07612efe4-kube-api-access-gptm8\") pod \"redhat-marketplace-bgkll\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.177677 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-utilities\") pod \"redhat-marketplace-bgkll\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.178358 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-catalog-content\") pod \"redhat-marketplace-bgkll\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.208757 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gptm8\" (UniqueName: \"kubernetes.io/projected/3870ac16-6934-481b-b91f-dea07612efe4-kube-api-access-gptm8\") pod \"redhat-marketplace-bgkll\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.241272 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.575278 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" event={"ID":"8e5451e1-267f-4401-8d38-e02d8a6f8191","Type":"ContainerStarted","Data":"b6d53cabc2f3b646a0a6ce9b2459bafd680d22a476134ec29221a6b522fb4f1e"} Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.575807 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" event={"ID":"8e5451e1-267f-4401-8d38-e02d8a6f8191","Type":"ContainerStarted","Data":"1f506c7b17617856511b55244cf69dfbafbb5937b04753cb1a8aa3a1430dbefe"} Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.862039 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" podStartSLOduration=1.8620200470000001 podStartE2EDuration="1.862020047s" podCreationTimestamp="2025-12-17 11:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 11:00:01.624410382 +0000 UTC m=+9537.169480334" watchObservedRunningTime="2025-12-17 11:00:01.862020047 +0000 UTC m=+9537.407089989" Dec 17 11:00:01 crc kubenswrapper[4966]: I1217 11:00:01.867837 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgkll"] Dec 17 11:00:01 crc kubenswrapper[4966]: W1217 11:00:01.888086 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3870ac16_6934_481b_b91f_dea07612efe4.slice/crio-a7fc1b3031ff966618a5ab4bf8b3db2d1089cb9dd0cd4c8c2db179a471f4daba WatchSource:0}: Error finding container a7fc1b3031ff966618a5ab4bf8b3db2d1089cb9dd0cd4c8c2db179a471f4daba: Status 404 returned error can't find the container with id a7fc1b3031ff966618a5ab4bf8b3db2d1089cb9dd0cd4c8c2db179a471f4daba Dec 17 11:00:02 crc kubenswrapper[4966]: I1217 11:00:02.585743 4966 generic.go:334] "Generic (PLEG): container finished" podID="3870ac16-6934-481b-b91f-dea07612efe4" containerID="2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446" exitCode=0 Dec 17 11:00:02 crc kubenswrapper[4966]: I1217 11:00:02.585813 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgkll" event={"ID":"3870ac16-6934-481b-b91f-dea07612efe4","Type":"ContainerDied","Data":"2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446"} Dec 17 11:00:02 crc kubenswrapper[4966]: I1217 11:00:02.585892 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgkll" event={"ID":"3870ac16-6934-481b-b91f-dea07612efe4","Type":"ContainerStarted","Data":"a7fc1b3031ff966618a5ab4bf8b3db2d1089cb9dd0cd4c8c2db179a471f4daba"} Dec 17 11:00:02 crc kubenswrapper[4966]: I1217 11:00:02.587392 4966 generic.go:334] "Generic (PLEG): container finished" podID="8e5451e1-267f-4401-8d38-e02d8a6f8191" containerID="b6d53cabc2f3b646a0a6ce9b2459bafd680d22a476134ec29221a6b522fb4f1e" exitCode=0 Dec 17 11:00:02 crc kubenswrapper[4966]: I1217 11:00:02.587434 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" event={"ID":"8e5451e1-267f-4401-8d38-e02d8a6f8191","Type":"ContainerDied","Data":"b6d53cabc2f3b646a0a6ce9b2459bafd680d22a476134ec29221a6b522fb4f1e"} Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.078533 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.257650 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2s6d\" (UniqueName: \"kubernetes.io/projected/8e5451e1-267f-4401-8d38-e02d8a6f8191-kube-api-access-t2s6d\") pod \"8e5451e1-267f-4401-8d38-e02d8a6f8191\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.257727 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e5451e1-267f-4401-8d38-e02d8a6f8191-secret-volume\") pod \"8e5451e1-267f-4401-8d38-e02d8a6f8191\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.257769 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e5451e1-267f-4401-8d38-e02d8a6f8191-config-volume\") pod \"8e5451e1-267f-4401-8d38-e02d8a6f8191\" (UID: \"8e5451e1-267f-4401-8d38-e02d8a6f8191\") " Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.258555 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e5451e1-267f-4401-8d38-e02d8a6f8191-config-volume" (OuterVolumeSpecName: "config-volume") pod "8e5451e1-267f-4401-8d38-e02d8a6f8191" (UID: "8e5451e1-267f-4401-8d38-e02d8a6f8191"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.264063 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e5451e1-267f-4401-8d38-e02d8a6f8191-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8e5451e1-267f-4401-8d38-e02d8a6f8191" (UID: "8e5451e1-267f-4401-8d38-e02d8a6f8191"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.264434 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e5451e1-267f-4401-8d38-e02d8a6f8191-kube-api-access-t2s6d" (OuterVolumeSpecName: "kube-api-access-t2s6d") pod "8e5451e1-267f-4401-8d38-e02d8a6f8191" (UID: "8e5451e1-267f-4401-8d38-e02d8a6f8191"). InnerVolumeSpecName "kube-api-access-t2s6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.360211 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2s6d\" (UniqueName: \"kubernetes.io/projected/8e5451e1-267f-4401-8d38-e02d8a6f8191-kube-api-access-t2s6d\") on node \"crc\" DevicePath \"\"" Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.360243 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8e5451e1-267f-4401-8d38-e02d8a6f8191-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.360252 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8e5451e1-267f-4401-8d38-e02d8a6f8191-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.608036 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" event={"ID":"8e5451e1-267f-4401-8d38-e02d8a6f8191","Type":"ContainerDied","Data":"1f506c7b17617856511b55244cf69dfbafbb5937b04753cb1a8aa3a1430dbefe"} Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.608253 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f506c7b17617856511b55244cf69dfbafbb5937b04753cb1a8aa3a1430dbefe" Dec 17 11:00:04 crc kubenswrapper[4966]: I1217 11:00:04.608150 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432820-wxk7c" Dec 17 11:00:05 crc kubenswrapper[4966]: I1217 11:00:05.174507 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx"] Dec 17 11:00:05 crc kubenswrapper[4966]: I1217 11:00:05.184376 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432775-9c6xx"] Dec 17 11:00:05 crc kubenswrapper[4966]: I1217 11:00:05.618118 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgkll" event={"ID":"3870ac16-6934-481b-b91f-dea07612efe4","Type":"ContainerStarted","Data":"f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a"} Dec 17 11:00:06 crc kubenswrapper[4966]: I1217 11:00:06.630052 4966 generic.go:334] "Generic (PLEG): container finished" podID="3870ac16-6934-481b-b91f-dea07612efe4" containerID="f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a" exitCode=0 Dec 17 11:00:06 crc kubenswrapper[4966]: I1217 11:00:06.630136 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgkll" event={"ID":"3870ac16-6934-481b-b91f-dea07612efe4","Type":"ContainerDied","Data":"f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a"} Dec 17 11:00:06 crc kubenswrapper[4966]: I1217 11:00:06.846643 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1199e280-30bf-4229-9b51-33172ca23479" path="/var/lib/kubelet/pods/1199e280-30bf-4229-9b51-33172ca23479/volumes" Dec 17 11:00:07 crc kubenswrapper[4966]: I1217 11:00:07.641191 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgkll" event={"ID":"3870ac16-6934-481b-b91f-dea07612efe4","Type":"ContainerStarted","Data":"52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4"} Dec 17 11:00:07 crc kubenswrapper[4966]: I1217 11:00:07.665207 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bgkll" podStartSLOduration=3.135637655 podStartE2EDuration="7.665184991s" podCreationTimestamp="2025-12-17 11:00:00 +0000 UTC" firstStartedPulling="2025-12-17 11:00:02.588013742 +0000 UTC m=+9538.133083684" lastFinishedPulling="2025-12-17 11:00:07.117561078 +0000 UTC m=+9542.662631020" observedRunningTime="2025-12-17 11:00:07.658426797 +0000 UTC m=+9543.203496749" watchObservedRunningTime="2025-12-17 11:00:07.665184991 +0000 UTC m=+9543.210254933" Dec 17 11:00:11 crc kubenswrapper[4966]: I1217 11:00:11.242003 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:11 crc kubenswrapper[4966]: I1217 11:00:11.242434 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:12 crc kubenswrapper[4966]: I1217 11:00:12.296637 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-bgkll" podUID="3870ac16-6934-481b-b91f-dea07612efe4" containerName="registry-server" probeResult="failure" output=< Dec 17 11:00:12 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 11:00:12 crc kubenswrapper[4966]: > Dec 17 11:00:16 crc kubenswrapper[4966]: I1217 11:00:16.807415 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 11:00:16 crc kubenswrapper[4966]: I1217 11:00:16.808985 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 11:00:21 crc kubenswrapper[4966]: I1217 11:00:21.316027 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:21 crc kubenswrapper[4966]: I1217 11:00:21.394448 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:21 crc kubenswrapper[4966]: I1217 11:00:21.563896 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgkll"] Dec 17 11:00:22 crc kubenswrapper[4966]: I1217 11:00:22.829196 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bgkll" podUID="3870ac16-6934-481b-b91f-dea07612efe4" containerName="registry-server" containerID="cri-o://52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4" gracePeriod=2 Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.368085 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.484179 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gptm8\" (UniqueName: \"kubernetes.io/projected/3870ac16-6934-481b-b91f-dea07612efe4-kube-api-access-gptm8\") pod \"3870ac16-6934-481b-b91f-dea07612efe4\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.484568 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-utilities\") pod \"3870ac16-6934-481b-b91f-dea07612efe4\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.484947 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-catalog-content\") pod \"3870ac16-6934-481b-b91f-dea07612efe4\" (UID: \"3870ac16-6934-481b-b91f-dea07612efe4\") " Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.485207 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-utilities" (OuterVolumeSpecName: "utilities") pod "3870ac16-6934-481b-b91f-dea07612efe4" (UID: "3870ac16-6934-481b-b91f-dea07612efe4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.486127 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.493557 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3870ac16-6934-481b-b91f-dea07612efe4-kube-api-access-gptm8" (OuterVolumeSpecName: "kube-api-access-gptm8") pod "3870ac16-6934-481b-b91f-dea07612efe4" (UID: "3870ac16-6934-481b-b91f-dea07612efe4"). InnerVolumeSpecName "kube-api-access-gptm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.509382 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3870ac16-6934-481b-b91f-dea07612efe4" (UID: "3870ac16-6934-481b-b91f-dea07612efe4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.587613 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gptm8\" (UniqueName: \"kubernetes.io/projected/3870ac16-6934-481b-b91f-dea07612efe4-kube-api-access-gptm8\") on node \"crc\" DevicePath \"\"" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.587653 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3870ac16-6934-481b-b91f-dea07612efe4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.840529 4966 generic.go:334] "Generic (PLEG): container finished" podID="3870ac16-6934-481b-b91f-dea07612efe4" containerID="52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4" exitCode=0 Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.840581 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgkll" event={"ID":"3870ac16-6934-481b-b91f-dea07612efe4","Type":"ContainerDied","Data":"52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4"} Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.840612 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgkll" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.840618 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgkll" event={"ID":"3870ac16-6934-481b-b91f-dea07612efe4","Type":"ContainerDied","Data":"a7fc1b3031ff966618a5ab4bf8b3db2d1089cb9dd0cd4c8c2db179a471f4daba"} Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.841435 4966 scope.go:117] "RemoveContainer" containerID="52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.876429 4966 scope.go:117] "RemoveContainer" containerID="f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.888506 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgkll"] Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.905530 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgkll"] Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.915808 4966 scope.go:117] "RemoveContainer" containerID="2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.968559 4966 scope.go:117] "RemoveContainer" containerID="52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4" Dec 17 11:00:23 crc kubenswrapper[4966]: E1217 11:00:23.969011 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4\": container with ID starting with 52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4 not found: ID does not exist" containerID="52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.969038 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4"} err="failed to get container status \"52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4\": rpc error: code = NotFound desc = could not find container \"52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4\": container with ID starting with 52699b5be64ae9db294cdbe83515f36dc7e9b5a7f91fad56d4fee63674b011c4 not found: ID does not exist" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.969060 4966 scope.go:117] "RemoveContainer" containerID="f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a" Dec 17 11:00:23 crc kubenswrapper[4966]: E1217 11:00:23.969357 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a\": container with ID starting with f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a not found: ID does not exist" containerID="f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.969375 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a"} err="failed to get container status \"f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a\": rpc error: code = NotFound desc = could not find container \"f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a\": container with ID starting with f56cebabfb6693e56ab99f6d9511c96a840586dd2f1eb3097c873f4122dcee3a not found: ID does not exist" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.969387 4966 scope.go:117] "RemoveContainer" containerID="2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446" Dec 17 11:00:23 crc kubenswrapper[4966]: E1217 11:00:23.969578 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446\": container with ID starting with 2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446 not found: ID does not exist" containerID="2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446" Dec 17 11:00:23 crc kubenswrapper[4966]: I1217 11:00:23.969596 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446"} err="failed to get container status \"2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446\": rpc error: code = NotFound desc = could not find container \"2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446\": container with ID starting with 2ccb717c4daf20931f4fc2db56cc6d5669a74f38f3ba54002b2028589ed67446 not found: ID does not exist" Dec 17 11:00:24 crc kubenswrapper[4966]: I1217 11:00:24.843337 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3870ac16-6934-481b-b91f-dea07612efe4" path="/var/lib/kubelet/pods/3870ac16-6934-481b-b91f-dea07612efe4/volumes" Dec 17 11:00:46 crc kubenswrapper[4966]: I1217 11:00:46.808006 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 11:00:46 crc kubenswrapper[4966]: I1217 11:00:46.808568 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.174021 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29432821-knzx7"] Dec 17 11:01:00 crc kubenswrapper[4966]: E1217 11:01:00.177227 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3870ac16-6934-481b-b91f-dea07612efe4" containerName="extract-utilities" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.177384 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3870ac16-6934-481b-b91f-dea07612efe4" containerName="extract-utilities" Dec 17 11:01:00 crc kubenswrapper[4966]: E1217 11:01:00.177407 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e5451e1-267f-4401-8d38-e02d8a6f8191" containerName="collect-profiles" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.177416 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e5451e1-267f-4401-8d38-e02d8a6f8191" containerName="collect-profiles" Dec 17 11:01:00 crc kubenswrapper[4966]: E1217 11:01:00.177428 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3870ac16-6934-481b-b91f-dea07612efe4" containerName="registry-server" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.177434 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3870ac16-6934-481b-b91f-dea07612efe4" containerName="registry-server" Dec 17 11:01:00 crc kubenswrapper[4966]: E1217 11:01:00.177444 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3870ac16-6934-481b-b91f-dea07612efe4" containerName="extract-content" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.177450 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="3870ac16-6934-481b-b91f-dea07612efe4" containerName="extract-content" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.177946 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e5451e1-267f-4401-8d38-e02d8a6f8191" containerName="collect-profiles" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.177963 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="3870ac16-6934-481b-b91f-dea07612efe4" containerName="registry-server" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.180073 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.218890 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29432821-knzx7"] Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.251475 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-config-data\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.251730 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-fernet-keys\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.251949 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-combined-ca-bundle\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.252063 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft5gd\" (UniqueName: \"kubernetes.io/projected/ad613eeb-41bd-46df-a76a-e71674083d9f-kube-api-access-ft5gd\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.353661 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-config-data\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.353706 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-fernet-keys\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.353782 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-combined-ca-bundle\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.353808 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft5gd\" (UniqueName: \"kubernetes.io/projected/ad613eeb-41bd-46df-a76a-e71674083d9f-kube-api-access-ft5gd\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.360905 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-fernet-keys\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.361806 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-combined-ca-bundle\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.367133 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-config-data\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.374457 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft5gd\" (UniqueName: \"kubernetes.io/projected/ad613eeb-41bd-46df-a76a-e71674083d9f-kube-api-access-ft5gd\") pod \"keystone-cron-29432821-knzx7\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.501062 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:00 crc kubenswrapper[4966]: I1217 11:01:00.971286 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29432821-knzx7"] Dec 17 11:01:01 crc kubenswrapper[4966]: I1217 11:01:01.199726 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432821-knzx7" event={"ID":"ad613eeb-41bd-46df-a76a-e71674083d9f","Type":"ContainerStarted","Data":"c2df846d2ea37d4aeb77a9b84101f58687629de95a21d0ca0f703e19aaf8780c"} Dec 17 11:01:01 crc kubenswrapper[4966]: I1217 11:01:01.199768 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432821-knzx7" event={"ID":"ad613eeb-41bd-46df-a76a-e71674083d9f","Type":"ContainerStarted","Data":"4d9bf351b0cd71251fc55fd836e6d60c5c2e3d65045cdee9fce41cb3fb972576"} Dec 17 11:01:01 crc kubenswrapper[4966]: I1217 11:01:01.224793 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29432821-knzx7" podStartSLOduration=1.224775716 podStartE2EDuration="1.224775716s" podCreationTimestamp="2025-12-17 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 11:01:01.224352044 +0000 UTC m=+9596.769422006" watchObservedRunningTime="2025-12-17 11:01:01.224775716 +0000 UTC m=+9596.769845658" Dec 17 11:01:04 crc kubenswrapper[4966]: I1217 11:01:04.239778 4966 generic.go:334] "Generic (PLEG): container finished" podID="ad613eeb-41bd-46df-a76a-e71674083d9f" containerID="c2df846d2ea37d4aeb77a9b84101f58687629de95a21d0ca0f703e19aaf8780c" exitCode=0 Dec 17 11:01:04 crc kubenswrapper[4966]: I1217 11:01:04.239884 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432821-knzx7" event={"ID":"ad613eeb-41bd-46df-a76a-e71674083d9f","Type":"ContainerDied","Data":"c2df846d2ea37d4aeb77a9b84101f58687629de95a21d0ca0f703e19aaf8780c"} Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.678527 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.756009 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-config-data\") pod \"ad613eeb-41bd-46df-a76a-e71674083d9f\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.756219 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-combined-ca-bundle\") pod \"ad613eeb-41bd-46df-a76a-e71674083d9f\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.756303 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-fernet-keys\") pod \"ad613eeb-41bd-46df-a76a-e71674083d9f\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.756365 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft5gd\" (UniqueName: \"kubernetes.io/projected/ad613eeb-41bd-46df-a76a-e71674083d9f-kube-api-access-ft5gd\") pod \"ad613eeb-41bd-46df-a76a-e71674083d9f\" (UID: \"ad613eeb-41bd-46df-a76a-e71674083d9f\") " Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.796642 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ad613eeb-41bd-46df-a76a-e71674083d9f" (UID: "ad613eeb-41bd-46df-a76a-e71674083d9f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.796651 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad613eeb-41bd-46df-a76a-e71674083d9f-kube-api-access-ft5gd" (OuterVolumeSpecName: "kube-api-access-ft5gd") pod "ad613eeb-41bd-46df-a76a-e71674083d9f" (UID: "ad613eeb-41bd-46df-a76a-e71674083d9f"). InnerVolumeSpecName "kube-api-access-ft5gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.800686 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad613eeb-41bd-46df-a76a-e71674083d9f" (UID: "ad613eeb-41bd-46df-a76a-e71674083d9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.824372 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-config-data" (OuterVolumeSpecName: "config-data") pod "ad613eeb-41bd-46df-a76a-e71674083d9f" (UID: "ad613eeb-41bd-46df-a76a-e71674083d9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.859107 4966 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.859141 4966 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.859154 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft5gd\" (UniqueName: \"kubernetes.io/projected/ad613eeb-41bd-46df-a76a-e71674083d9f-kube-api-access-ft5gd\") on node \"crc\" DevicePath \"\"" Dec 17 11:01:05 crc kubenswrapper[4966]: I1217 11:01:05.859168 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad613eeb-41bd-46df-a76a-e71674083d9f-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 11:01:06 crc kubenswrapper[4966]: I1217 11:01:06.259639 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29432821-knzx7" event={"ID":"ad613eeb-41bd-46df-a76a-e71674083d9f","Type":"ContainerDied","Data":"4d9bf351b0cd71251fc55fd836e6d60c5c2e3d65045cdee9fce41cb3fb972576"} Dec 17 11:01:06 crc kubenswrapper[4966]: I1217 11:01:06.259688 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d9bf351b0cd71251fc55fd836e6d60c5c2e3d65045cdee9fce41cb3fb972576" Dec 17 11:01:06 crc kubenswrapper[4966]: I1217 11:01:06.259725 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29432821-knzx7" Dec 17 11:01:06 crc kubenswrapper[4966]: I1217 11:01:06.380674 4966 scope.go:117] "RemoveContainer" containerID="22ed0010f84ded624d7c25baa8623747c767714da71173e87eba5ef6160368e1" Dec 17 11:01:16 crc kubenswrapper[4966]: I1217 11:01:16.808312 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 11:01:16 crc kubenswrapper[4966]: I1217 11:01:16.809665 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 11:01:16 crc kubenswrapper[4966]: I1217 11:01:16.809781 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 11:01:16 crc kubenswrapper[4966]: I1217 11:01:16.810700 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 11:01:16 crc kubenswrapper[4966]: I1217 11:01:16.810840 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" gracePeriod=600 Dec 17 11:01:17 crc kubenswrapper[4966]: E1217 11:01:17.142763 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:01:17 crc kubenswrapper[4966]: I1217 11:01:17.387527 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" exitCode=0 Dec 17 11:01:17 crc kubenswrapper[4966]: I1217 11:01:17.387571 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b"} Dec 17 11:01:17 crc kubenswrapper[4966]: I1217 11:01:17.387601 4966 scope.go:117] "RemoveContainer" containerID="05f4c1c9032aed889945ebfbf97186e7dfeda826202a490e0f1d86c1a437ed61" Dec 17 11:01:17 crc kubenswrapper[4966]: I1217 11:01:17.391438 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:01:17 crc kubenswrapper[4966]: E1217 11:01:17.391706 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:01:28 crc kubenswrapper[4966]: I1217 11:01:28.831995 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:01:28 crc kubenswrapper[4966]: E1217 11:01:28.833081 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:01:39 crc kubenswrapper[4966]: I1217 11:01:39.831377 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:01:39 crc kubenswrapper[4966]: E1217 11:01:39.834355 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:01:53 crc kubenswrapper[4966]: I1217 11:01:53.830476 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:01:53 crc kubenswrapper[4966]: E1217 11:01:53.831179 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:02:08 crc kubenswrapper[4966]: I1217 11:02:08.830411 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:02:08 crc kubenswrapper[4966]: E1217 11:02:08.831266 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:02:17 crc kubenswrapper[4966]: I1217 11:02:17.172140 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-lcqdh" podUID="46dfa510-96af-4d45-9be7-5bd9a8588a61" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 17 11:02:19 crc kubenswrapper[4966]: I1217 11:02:19.830472 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:02:19 crc kubenswrapper[4966]: E1217 11:02:19.831013 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:02:33 crc kubenswrapper[4966]: I1217 11:02:33.830141 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:02:33 crc kubenswrapper[4966]: E1217 11:02:33.830996 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.291127 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xs7fz"] Dec 17 11:02:37 crc kubenswrapper[4966]: E1217 11:02:37.292378 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad613eeb-41bd-46df-a76a-e71674083d9f" containerName="keystone-cron" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.292394 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad613eeb-41bd-46df-a76a-e71674083d9f" containerName="keystone-cron" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.293451 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad613eeb-41bd-46df-a76a-e71674083d9f" containerName="keystone-cron" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.298742 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.352266 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xs7fz"] Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.413269 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-utilities\") pod \"community-operators-xs7fz\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.414282 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-catalog-content\") pod \"community-operators-xs7fz\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.414535 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2fxc\" (UniqueName: \"kubernetes.io/projected/71862aed-171f-4aea-bb63-2518e329dedc-kube-api-access-l2fxc\") pod \"community-operators-xs7fz\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.518229 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-utilities\") pod \"community-operators-xs7fz\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.518338 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-catalog-content\") pod \"community-operators-xs7fz\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.518386 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2fxc\" (UniqueName: \"kubernetes.io/projected/71862aed-171f-4aea-bb63-2518e329dedc-kube-api-access-l2fxc\") pod \"community-operators-xs7fz\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.518932 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-catalog-content\") pod \"community-operators-xs7fz\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.518972 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-utilities\") pod \"community-operators-xs7fz\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.542969 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2fxc\" (UniqueName: \"kubernetes.io/projected/71862aed-171f-4aea-bb63-2518e329dedc-kube-api-access-l2fxc\") pod \"community-operators-xs7fz\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:37 crc kubenswrapper[4966]: I1217 11:02:37.632427 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:38 crc kubenswrapper[4966]: I1217 11:02:38.166466 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xs7fz"] Dec 17 11:02:38 crc kubenswrapper[4966]: I1217 11:02:38.396980 4966 generic.go:334] "Generic (PLEG): container finished" podID="71862aed-171f-4aea-bb63-2518e329dedc" containerID="4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4" exitCode=0 Dec 17 11:02:38 crc kubenswrapper[4966]: I1217 11:02:38.397031 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xs7fz" event={"ID":"71862aed-171f-4aea-bb63-2518e329dedc","Type":"ContainerDied","Data":"4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4"} Dec 17 11:02:38 crc kubenswrapper[4966]: I1217 11:02:38.397061 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xs7fz" event={"ID":"71862aed-171f-4aea-bb63-2518e329dedc","Type":"ContainerStarted","Data":"48b2d5fc091b267311bf75e92e74f1fa16bdbb81d235704628f483a38bbad406"} Dec 17 11:02:38 crc kubenswrapper[4966]: I1217 11:02:38.400055 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 11:02:40 crc kubenswrapper[4966]: I1217 11:02:40.416473 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xs7fz" event={"ID":"71862aed-171f-4aea-bb63-2518e329dedc","Type":"ContainerStarted","Data":"f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714"} Dec 17 11:02:41 crc kubenswrapper[4966]: I1217 11:02:41.436782 4966 generic.go:334] "Generic (PLEG): container finished" podID="71862aed-171f-4aea-bb63-2518e329dedc" containerID="f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714" exitCode=0 Dec 17 11:02:41 crc kubenswrapper[4966]: I1217 11:02:41.436836 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xs7fz" event={"ID":"71862aed-171f-4aea-bb63-2518e329dedc","Type":"ContainerDied","Data":"f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714"} Dec 17 11:02:42 crc kubenswrapper[4966]: I1217 11:02:42.448092 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xs7fz" event={"ID":"71862aed-171f-4aea-bb63-2518e329dedc","Type":"ContainerStarted","Data":"b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e"} Dec 17 11:02:42 crc kubenswrapper[4966]: I1217 11:02:42.477401 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xs7fz" podStartSLOduration=1.944754345 podStartE2EDuration="5.477380503s" podCreationTimestamp="2025-12-17 11:02:37 +0000 UTC" firstStartedPulling="2025-12-17 11:02:38.398689744 +0000 UTC m=+9693.943759696" lastFinishedPulling="2025-12-17 11:02:41.931315892 +0000 UTC m=+9697.476385854" observedRunningTime="2025-12-17 11:02:42.475992145 +0000 UTC m=+9698.021062097" watchObservedRunningTime="2025-12-17 11:02:42.477380503 +0000 UTC m=+9698.022450445" Dec 17 11:02:46 crc kubenswrapper[4966]: I1217 11:02:46.833847 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:02:46 crc kubenswrapper[4966]: E1217 11:02:46.834651 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:02:47 crc kubenswrapper[4966]: I1217 11:02:47.633119 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:47 crc kubenswrapper[4966]: I1217 11:02:47.633184 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:47 crc kubenswrapper[4966]: I1217 11:02:47.686632 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:48 crc kubenswrapper[4966]: I1217 11:02:48.571511 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:48 crc kubenswrapper[4966]: I1217 11:02:48.640204 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xs7fz"] Dec 17 11:02:50 crc kubenswrapper[4966]: I1217 11:02:50.534761 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xs7fz" podUID="71862aed-171f-4aea-bb63-2518e329dedc" containerName="registry-server" containerID="cri-o://b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e" gracePeriod=2 Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.081835 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.092430 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-catalog-content\") pod \"71862aed-171f-4aea-bb63-2518e329dedc\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.092625 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-utilities\") pod \"71862aed-171f-4aea-bb63-2518e329dedc\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.092654 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2fxc\" (UniqueName: \"kubernetes.io/projected/71862aed-171f-4aea-bb63-2518e329dedc-kube-api-access-l2fxc\") pod \"71862aed-171f-4aea-bb63-2518e329dedc\" (UID: \"71862aed-171f-4aea-bb63-2518e329dedc\") " Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.093802 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-utilities" (OuterVolumeSpecName: "utilities") pod "71862aed-171f-4aea-bb63-2518e329dedc" (UID: "71862aed-171f-4aea-bb63-2518e329dedc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.100002 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71862aed-171f-4aea-bb63-2518e329dedc-kube-api-access-l2fxc" (OuterVolumeSpecName: "kube-api-access-l2fxc") pod "71862aed-171f-4aea-bb63-2518e329dedc" (UID: "71862aed-171f-4aea-bb63-2518e329dedc"). InnerVolumeSpecName "kube-api-access-l2fxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.153482 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71862aed-171f-4aea-bb63-2518e329dedc" (UID: "71862aed-171f-4aea-bb63-2518e329dedc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.194229 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.194268 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71862aed-171f-4aea-bb63-2518e329dedc-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.194282 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2fxc\" (UniqueName: \"kubernetes.io/projected/71862aed-171f-4aea-bb63-2518e329dedc-kube-api-access-l2fxc\") on node \"crc\" DevicePath \"\"" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.544321 4966 generic.go:334] "Generic (PLEG): container finished" podID="71862aed-171f-4aea-bb63-2518e329dedc" containerID="b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e" exitCode=0 Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.544383 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xs7fz" event={"ID":"71862aed-171f-4aea-bb63-2518e329dedc","Type":"ContainerDied","Data":"b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e"} Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.544391 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xs7fz" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.544416 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xs7fz" event={"ID":"71862aed-171f-4aea-bb63-2518e329dedc","Type":"ContainerDied","Data":"48b2d5fc091b267311bf75e92e74f1fa16bdbb81d235704628f483a38bbad406"} Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.544445 4966 scope.go:117] "RemoveContainer" containerID="b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.578944 4966 scope.go:117] "RemoveContainer" containerID="f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.594185 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xs7fz"] Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.602855 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xs7fz"] Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.607340 4966 scope.go:117] "RemoveContainer" containerID="4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.663393 4966 scope.go:117] "RemoveContainer" containerID="b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e" Dec 17 11:02:51 crc kubenswrapper[4966]: E1217 11:02:51.669218 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e\": container with ID starting with b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e not found: ID does not exist" containerID="b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.669446 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e"} err="failed to get container status \"b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e\": rpc error: code = NotFound desc = could not find container \"b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e\": container with ID starting with b29c363229a6e320318ca310fb53f9e755b4c93276de4d1318f2c8034531ae0e not found: ID does not exist" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.669480 4966 scope.go:117] "RemoveContainer" containerID="f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714" Dec 17 11:02:51 crc kubenswrapper[4966]: E1217 11:02:51.670402 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714\": container with ID starting with f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714 not found: ID does not exist" containerID="f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.670442 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714"} err="failed to get container status \"f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714\": rpc error: code = NotFound desc = could not find container \"f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714\": container with ID starting with f051b03fd74ac8f382907e7ec9cf9abca9b5c771c8c2cb0136555747611a9714 not found: ID does not exist" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.670468 4966 scope.go:117] "RemoveContainer" containerID="4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4" Dec 17 11:02:51 crc kubenswrapper[4966]: E1217 11:02:51.670865 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4\": container with ID starting with 4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4 not found: ID does not exist" containerID="4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4" Dec 17 11:02:51 crc kubenswrapper[4966]: I1217 11:02:51.671015 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4"} err="failed to get container status \"4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4\": rpc error: code = NotFound desc = could not find container \"4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4\": container with ID starting with 4dc7cc0404ca960c256d18ab29c2244c17c9af76837d74100c3d0aa56b9faed4 not found: ID does not exist" Dec 17 11:02:52 crc kubenswrapper[4966]: I1217 11:02:52.841863 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71862aed-171f-4aea-bb63-2518e329dedc" path="/var/lib/kubelet/pods/71862aed-171f-4aea-bb63-2518e329dedc/volumes" Dec 17 11:02:58 crc kubenswrapper[4966]: I1217 11:02:58.830087 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:02:58 crc kubenswrapper[4966]: E1217 11:02:58.831766 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:03:09 crc kubenswrapper[4966]: I1217 11:03:09.830627 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:03:09 crc kubenswrapper[4966]: E1217 11:03:09.831862 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:03:20 crc kubenswrapper[4966]: I1217 11:03:20.830430 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:03:20 crc kubenswrapper[4966]: E1217 11:03:20.831286 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:03:35 crc kubenswrapper[4966]: I1217 11:03:35.831375 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:03:35 crc kubenswrapper[4966]: E1217 11:03:35.832171 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:03:48 crc kubenswrapper[4966]: I1217 11:03:48.830655 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:03:48 crc kubenswrapper[4966]: E1217 11:03:48.831541 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:04:00 crc kubenswrapper[4966]: I1217 11:04:00.831863 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:04:00 crc kubenswrapper[4966]: E1217 11:04:00.833510 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:04:15 crc kubenswrapper[4966]: I1217 11:04:15.830570 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:04:15 crc kubenswrapper[4966]: E1217 11:04:15.832409 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:04:28 crc kubenswrapper[4966]: I1217 11:04:28.830797 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:04:28 crc kubenswrapper[4966]: E1217 11:04:28.831513 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:04:40 crc kubenswrapper[4966]: I1217 11:04:40.830768 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:04:40 crc kubenswrapper[4966]: E1217 11:04:40.831627 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:04:55 crc kubenswrapper[4966]: I1217 11:04:55.831073 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:04:55 crc kubenswrapper[4966]: E1217 11:04:55.831895 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:05:07 crc kubenswrapper[4966]: I1217 11:05:07.830506 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:05:07 crc kubenswrapper[4966]: E1217 11:05:07.831441 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:05:22 crc kubenswrapper[4966]: I1217 11:05:22.830378 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:05:22 crc kubenswrapper[4966]: E1217 11:05:22.831061 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:05:35 crc kubenswrapper[4966]: I1217 11:05:35.832950 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:05:35 crc kubenswrapper[4966]: E1217 11:05:35.834160 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:05:50 crc kubenswrapper[4966]: I1217 11:05:50.830467 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:05:50 crc kubenswrapper[4966]: E1217 11:05:50.831196 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:06:01 crc kubenswrapper[4966]: I1217 11:06:01.830651 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:06:01 crc kubenswrapper[4966]: E1217 11:06:01.831296 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:06:13 crc kubenswrapper[4966]: I1217 11:06:13.830661 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:06:13 crc kubenswrapper[4966]: E1217 11:06:13.831673 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:06:25 crc kubenswrapper[4966]: I1217 11:06:25.830316 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:06:26 crc kubenswrapper[4966]: I1217 11:06:26.621245 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"774bf964c7159826b3f4e46131c91b45d09332fd4b4923df8be5ab7d71835abb"} Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.289944 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zfrrm"] Dec 17 11:06:52 crc kubenswrapper[4966]: E1217 11:06:52.291016 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71862aed-171f-4aea-bb63-2518e329dedc" containerName="registry-server" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.291036 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="71862aed-171f-4aea-bb63-2518e329dedc" containerName="registry-server" Dec 17 11:06:52 crc kubenswrapper[4966]: E1217 11:06:52.291049 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71862aed-171f-4aea-bb63-2518e329dedc" containerName="extract-utilities" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.291057 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="71862aed-171f-4aea-bb63-2518e329dedc" containerName="extract-utilities" Dec 17 11:06:52 crc kubenswrapper[4966]: E1217 11:06:52.291081 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71862aed-171f-4aea-bb63-2518e329dedc" containerName="extract-content" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.291089 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="71862aed-171f-4aea-bb63-2518e329dedc" containerName="extract-content" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.291322 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="71862aed-171f-4aea-bb63-2518e329dedc" containerName="registry-server" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.293979 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.325647 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zfrrm"] Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.373985 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6cw5\" (UniqueName: \"kubernetes.io/projected/6fde326e-f683-4cc3-b90f-8d41f74541e6-kube-api-access-g6cw5\") pod \"certified-operators-zfrrm\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.374087 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-utilities\") pod \"certified-operators-zfrrm\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.374155 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-catalog-content\") pod \"certified-operators-zfrrm\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.476279 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-utilities\") pod \"certified-operators-zfrrm\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.476382 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-catalog-content\") pod \"certified-operators-zfrrm\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.476493 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6cw5\" (UniqueName: \"kubernetes.io/projected/6fde326e-f683-4cc3-b90f-8d41f74541e6-kube-api-access-g6cw5\") pod \"certified-operators-zfrrm\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.477010 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-utilities\") pod \"certified-operators-zfrrm\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.477131 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-catalog-content\") pod \"certified-operators-zfrrm\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.498359 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6cw5\" (UniqueName: \"kubernetes.io/projected/6fde326e-f683-4cc3-b90f-8d41f74541e6-kube-api-access-g6cw5\") pod \"certified-operators-zfrrm\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:52 crc kubenswrapper[4966]: I1217 11:06:52.672783 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:06:53 crc kubenswrapper[4966]: I1217 11:06:53.225747 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zfrrm"] Dec 17 11:06:53 crc kubenswrapper[4966]: I1217 11:06:53.884698 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfrrm" event={"ID":"6fde326e-f683-4cc3-b90f-8d41f74541e6","Type":"ContainerStarted","Data":"c56f9829e8844da7e08a06f8f35c2703ba74c00164f24eaa816e60d9f312a57d"} Dec 17 11:06:54 crc kubenswrapper[4966]: I1217 11:06:54.894475 4966 generic.go:334] "Generic (PLEG): container finished" podID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerID="ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d" exitCode=0 Dec 17 11:06:54 crc kubenswrapper[4966]: I1217 11:06:54.895305 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfrrm" event={"ID":"6fde326e-f683-4cc3-b90f-8d41f74541e6","Type":"ContainerDied","Data":"ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d"} Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.290334 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pz77w"] Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.293143 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.301996 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pz77w"] Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.431108 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-utilities\") pod \"redhat-operators-pz77w\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.431209 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w8gz\" (UniqueName: \"kubernetes.io/projected/e3fb435e-b61e-4aac-8629-9a07d685649e-kube-api-access-8w8gz\") pod \"redhat-operators-pz77w\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.431452 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-catalog-content\") pod \"redhat-operators-pz77w\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.533501 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-catalog-content\") pod \"redhat-operators-pz77w\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.533619 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-utilities\") pod \"redhat-operators-pz77w\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.533660 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w8gz\" (UniqueName: \"kubernetes.io/projected/e3fb435e-b61e-4aac-8629-9a07d685649e-kube-api-access-8w8gz\") pod \"redhat-operators-pz77w\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.534685 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-utilities\") pod \"redhat-operators-pz77w\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.534948 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-catalog-content\") pod \"redhat-operators-pz77w\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.600467 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w8gz\" (UniqueName: \"kubernetes.io/projected/e3fb435e-b61e-4aac-8629-9a07d685649e-kube-api-access-8w8gz\") pod \"redhat-operators-pz77w\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:55 crc kubenswrapper[4966]: I1217 11:06:55.646590 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:06:56 crc kubenswrapper[4966]: I1217 11:06:56.305312 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pz77w"] Dec 17 11:06:56 crc kubenswrapper[4966]: W1217 11:06:56.316462 4966 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3fb435e_b61e_4aac_8629_9a07d685649e.slice/crio-762d75517ba7adf73c8d52965f09da63e4a271aa5f1a8063934ca5e5359b20cf WatchSource:0}: Error finding container 762d75517ba7adf73c8d52965f09da63e4a271aa5f1a8063934ca5e5359b20cf: Status 404 returned error can't find the container with id 762d75517ba7adf73c8d52965f09da63e4a271aa5f1a8063934ca5e5359b20cf Dec 17 11:06:56 crc kubenswrapper[4966]: I1217 11:06:56.921064 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfrrm" event={"ID":"6fde326e-f683-4cc3-b90f-8d41f74541e6","Type":"ContainerStarted","Data":"96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece"} Dec 17 11:06:56 crc kubenswrapper[4966]: I1217 11:06:56.924586 4966 generic.go:334] "Generic (PLEG): container finished" podID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerID="3fb2fd5d34ef70654a4e4fca0d69685ce164773047e58a3103066a20e1e27a17" exitCode=0 Dec 17 11:06:56 crc kubenswrapper[4966]: I1217 11:06:56.924636 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz77w" event={"ID":"e3fb435e-b61e-4aac-8629-9a07d685649e","Type":"ContainerDied","Data":"3fb2fd5d34ef70654a4e4fca0d69685ce164773047e58a3103066a20e1e27a17"} Dec 17 11:06:56 crc kubenswrapper[4966]: I1217 11:06:56.924682 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz77w" event={"ID":"e3fb435e-b61e-4aac-8629-9a07d685649e","Type":"ContainerStarted","Data":"762d75517ba7adf73c8d52965f09da63e4a271aa5f1a8063934ca5e5359b20cf"} Dec 17 11:06:58 crc kubenswrapper[4966]: I1217 11:06:58.979188 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfrrm" event={"ID":"6fde326e-f683-4cc3-b90f-8d41f74541e6","Type":"ContainerDied","Data":"96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece"} Dec 17 11:06:58 crc kubenswrapper[4966]: I1217 11:06:58.979128 4966 generic.go:334] "Generic (PLEG): container finished" podID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerID="96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece" exitCode=0 Dec 17 11:06:58 crc kubenswrapper[4966]: I1217 11:06:58.993586 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz77w" event={"ID":"e3fb435e-b61e-4aac-8629-9a07d685649e","Type":"ContainerStarted","Data":"2c910ce0e269ba5f0e0fb85927cbdf76f031f8cbd3589905b4ef1d6e0c3f078f"} Dec 17 11:07:00 crc kubenswrapper[4966]: I1217 11:07:00.003812 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfrrm" event={"ID":"6fde326e-f683-4cc3-b90f-8d41f74541e6","Type":"ContainerStarted","Data":"9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7"} Dec 17 11:07:00 crc kubenswrapper[4966]: I1217 11:07:00.024653 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zfrrm" podStartSLOduration=3.519054592 podStartE2EDuration="8.024280176s" podCreationTimestamp="2025-12-17 11:06:52 +0000 UTC" firstStartedPulling="2025-12-17 11:06:54.898147261 +0000 UTC m=+9950.443217203" lastFinishedPulling="2025-12-17 11:06:59.403372845 +0000 UTC m=+9954.948442787" observedRunningTime="2025-12-17 11:07:00.018486197 +0000 UTC m=+9955.563556149" watchObservedRunningTime="2025-12-17 11:07:00.024280176 +0000 UTC m=+9955.569350118" Dec 17 11:07:02 crc kubenswrapper[4966]: I1217 11:07:02.673557 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:07:02 crc kubenswrapper[4966]: I1217 11:07:02.674198 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:07:03 crc kubenswrapper[4966]: I1217 11:07:03.034187 4966 generic.go:334] "Generic (PLEG): container finished" podID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerID="2c910ce0e269ba5f0e0fb85927cbdf76f031f8cbd3589905b4ef1d6e0c3f078f" exitCode=0 Dec 17 11:07:03 crc kubenswrapper[4966]: I1217 11:07:03.034275 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz77w" event={"ID":"e3fb435e-b61e-4aac-8629-9a07d685649e","Type":"ContainerDied","Data":"2c910ce0e269ba5f0e0fb85927cbdf76f031f8cbd3589905b4ef1d6e0c3f078f"} Dec 17 11:07:03 crc kubenswrapper[4966]: I1217 11:07:03.730134 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-zfrrm" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerName="registry-server" probeResult="failure" output=< Dec 17 11:07:03 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 11:07:03 crc kubenswrapper[4966]: > Dec 17 11:07:04 crc kubenswrapper[4966]: I1217 11:07:04.045814 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz77w" event={"ID":"e3fb435e-b61e-4aac-8629-9a07d685649e","Type":"ContainerStarted","Data":"14fbf698e9017f0cf0fa94eece7c4442e948396433ba3938f18f8d6abacf3053"} Dec 17 11:07:04 crc kubenswrapper[4966]: I1217 11:07:04.070169 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pz77w" podStartSLOduration=2.336472579 podStartE2EDuration="9.070148471s" podCreationTimestamp="2025-12-17 11:06:55 +0000 UTC" firstStartedPulling="2025-12-17 11:06:56.928854071 +0000 UTC m=+9952.473924013" lastFinishedPulling="2025-12-17 11:07:03.662529973 +0000 UTC m=+9959.207599905" observedRunningTime="2025-12-17 11:07:04.063531211 +0000 UTC m=+9959.608601153" watchObservedRunningTime="2025-12-17 11:07:04.070148471 +0000 UTC m=+9959.615218413" Dec 17 11:07:05 crc kubenswrapper[4966]: I1217 11:07:05.647117 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:07:05 crc kubenswrapper[4966]: I1217 11:07:05.647460 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:07:06 crc kubenswrapper[4966]: I1217 11:07:06.702278 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pz77w" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="registry-server" probeResult="failure" output=< Dec 17 11:07:06 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 11:07:06 crc kubenswrapper[4966]: > Dec 17 11:07:12 crc kubenswrapper[4966]: I1217 11:07:12.724334 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:07:12 crc kubenswrapper[4966]: I1217 11:07:12.780313 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:07:16 crc kubenswrapper[4966]: I1217 11:07:16.475546 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zfrrm"] Dec 17 11:07:16 crc kubenswrapper[4966]: I1217 11:07:16.477020 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zfrrm" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerName="registry-server" containerID="cri-o://9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7" gracePeriod=2 Dec 17 11:07:16 crc kubenswrapper[4966]: I1217 11:07:16.688852 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pz77w" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="registry-server" probeResult="failure" output=< Dec 17 11:07:16 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 11:07:16 crc kubenswrapper[4966]: > Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.253860 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.400384 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-utilities\") pod \"6fde326e-f683-4cc3-b90f-8d41f74541e6\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.400532 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6cw5\" (UniqueName: \"kubernetes.io/projected/6fde326e-f683-4cc3-b90f-8d41f74541e6-kube-api-access-g6cw5\") pod \"6fde326e-f683-4cc3-b90f-8d41f74541e6\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.400592 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-catalog-content\") pod \"6fde326e-f683-4cc3-b90f-8d41f74541e6\" (UID: \"6fde326e-f683-4cc3-b90f-8d41f74541e6\") " Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.402068 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-utilities" (OuterVolumeSpecName: "utilities") pod "6fde326e-f683-4cc3-b90f-8d41f74541e6" (UID: "6fde326e-f683-4cc3-b90f-8d41f74541e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.416182 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fde326e-f683-4cc3-b90f-8d41f74541e6-kube-api-access-g6cw5" (OuterVolumeSpecName: "kube-api-access-g6cw5") pod "6fde326e-f683-4cc3-b90f-8d41f74541e6" (UID: "6fde326e-f683-4cc3-b90f-8d41f74541e6"). InnerVolumeSpecName "kube-api-access-g6cw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.470040 4966 generic.go:334] "Generic (PLEG): container finished" podID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerID="9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7" exitCode=0 Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.470097 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfrrm" event={"ID":"6fde326e-f683-4cc3-b90f-8d41f74541e6","Type":"ContainerDied","Data":"9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7"} Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.470169 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zfrrm" event={"ID":"6fde326e-f683-4cc3-b90f-8d41f74541e6","Type":"ContainerDied","Data":"c56f9829e8844da7e08a06f8f35c2703ba74c00164f24eaa816e60d9f312a57d"} Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.470175 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zfrrm" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.470622 4966 scope.go:117] "RemoveContainer" containerID="9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.486502 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fde326e-f683-4cc3-b90f-8d41f74541e6" (UID: "6fde326e-f683-4cc3-b90f-8d41f74541e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.493371 4966 scope.go:117] "RemoveContainer" containerID="96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.503362 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6cw5\" (UniqueName: \"kubernetes.io/projected/6fde326e-f683-4cc3-b90f-8d41f74541e6-kube-api-access-g6cw5\") on node \"crc\" DevicePath \"\"" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.503390 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.503400 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fde326e-f683-4cc3-b90f-8d41f74541e6-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.525211 4966 scope.go:117] "RemoveContainer" containerID="ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.567430 4966 scope.go:117] "RemoveContainer" containerID="9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7" Dec 17 11:07:17 crc kubenswrapper[4966]: E1217 11:07:17.568757 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7\": container with ID starting with 9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7 not found: ID does not exist" containerID="9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.569112 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7"} err="failed to get container status \"9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7\": rpc error: code = NotFound desc = could not find container \"9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7\": container with ID starting with 9fade06df20be92150c91a5dfabe643203633196821b6262ba5205d88d4406c7 not found: ID does not exist" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.569154 4966 scope.go:117] "RemoveContainer" containerID="96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece" Dec 17 11:07:17 crc kubenswrapper[4966]: E1217 11:07:17.569478 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece\": container with ID starting with 96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece not found: ID does not exist" containerID="96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.569510 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece"} err="failed to get container status \"96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece\": rpc error: code = NotFound desc = could not find container \"96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece\": container with ID starting with 96c279ef34af486375b24ecd2d0c49dd33ca9800474ad07ca3206aa9f5c27ece not found: ID does not exist" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.569541 4966 scope.go:117] "RemoveContainer" containerID="ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d" Dec 17 11:07:17 crc kubenswrapper[4966]: E1217 11:07:17.569791 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d\": container with ID starting with ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d not found: ID does not exist" containerID="ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.569817 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d"} err="failed to get container status \"ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d\": rpc error: code = NotFound desc = could not find container \"ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d\": container with ID starting with ec2e4b4d5af5697478623ec060210f04f76c2c2bb7991ca1c7eb59b7bb95131d not found: ID does not exist" Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.809916 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zfrrm"] Dec 17 11:07:17 crc kubenswrapper[4966]: I1217 11:07:17.818780 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zfrrm"] Dec 17 11:07:18 crc kubenswrapper[4966]: I1217 11:07:18.842192 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" path="/var/lib/kubelet/pods/6fde326e-f683-4cc3-b90f-8d41f74541e6/volumes" Dec 17 11:07:25 crc kubenswrapper[4966]: I1217 11:07:25.704917 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:07:25 crc kubenswrapper[4966]: I1217 11:07:25.757389 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:07:25 crc kubenswrapper[4966]: I1217 11:07:25.944901 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pz77w"] Dec 17 11:07:27 crc kubenswrapper[4966]: I1217 11:07:27.575333 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pz77w" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="registry-server" containerID="cri-o://14fbf698e9017f0cf0fa94eece7c4442e948396433ba3938f18f8d6abacf3053" gracePeriod=2 Dec 17 11:07:28 crc kubenswrapper[4966]: I1217 11:07:28.607846 4966 generic.go:334] "Generic (PLEG): container finished" podID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerID="14fbf698e9017f0cf0fa94eece7c4442e948396433ba3938f18f8d6abacf3053" exitCode=0 Dec 17 11:07:28 crc kubenswrapper[4966]: I1217 11:07:28.608224 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz77w" event={"ID":"e3fb435e-b61e-4aac-8629-9a07d685649e","Type":"ContainerDied","Data":"14fbf698e9017f0cf0fa94eece7c4442e948396433ba3938f18f8d6abacf3053"} Dec 17 11:07:28 crc kubenswrapper[4966]: I1217 11:07:28.890851 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.069795 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-catalog-content\") pod \"e3fb435e-b61e-4aac-8629-9a07d685649e\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.074533 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-utilities\") pod \"e3fb435e-b61e-4aac-8629-9a07d685649e\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.074577 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w8gz\" (UniqueName: \"kubernetes.io/projected/e3fb435e-b61e-4aac-8629-9a07d685649e-kube-api-access-8w8gz\") pod \"e3fb435e-b61e-4aac-8629-9a07d685649e\" (UID: \"e3fb435e-b61e-4aac-8629-9a07d685649e\") " Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.074991 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-utilities" (OuterVolumeSpecName: "utilities") pod "e3fb435e-b61e-4aac-8629-9a07d685649e" (UID: "e3fb435e-b61e-4aac-8629-9a07d685649e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.075456 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.091639 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3fb435e-b61e-4aac-8629-9a07d685649e-kube-api-access-8w8gz" (OuterVolumeSpecName: "kube-api-access-8w8gz") pod "e3fb435e-b61e-4aac-8629-9a07d685649e" (UID: "e3fb435e-b61e-4aac-8629-9a07d685649e"). InnerVolumeSpecName "kube-api-access-8w8gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.163030 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3fb435e-b61e-4aac-8629-9a07d685649e" (UID: "e3fb435e-b61e-4aac-8629-9a07d685649e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.178059 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w8gz\" (UniqueName: \"kubernetes.io/projected/e3fb435e-b61e-4aac-8629-9a07d685649e-kube-api-access-8w8gz\") on node \"crc\" DevicePath \"\"" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.178101 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3fb435e-b61e-4aac-8629-9a07d685649e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.620021 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pz77w" event={"ID":"e3fb435e-b61e-4aac-8629-9a07d685649e","Type":"ContainerDied","Data":"762d75517ba7adf73c8d52965f09da63e4a271aa5f1a8063934ca5e5359b20cf"} Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.620099 4966 scope.go:117] "RemoveContainer" containerID="14fbf698e9017f0cf0fa94eece7c4442e948396433ba3938f18f8d6abacf3053" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.621212 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pz77w" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.658515 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pz77w"] Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.660543 4966 scope.go:117] "RemoveContainer" containerID="2c910ce0e269ba5f0e0fb85927cbdf76f031f8cbd3589905b4ef1d6e0c3f078f" Dec 17 11:07:29 crc kubenswrapper[4966]: I1217 11:07:29.668198 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pz77w"] Dec 17 11:07:30 crc kubenswrapper[4966]: I1217 11:07:30.119376 4966 scope.go:117] "RemoveContainer" containerID="3fb2fd5d34ef70654a4e4fca0d69685ce164773047e58a3103066a20e1e27a17" Dec 17 11:07:30 crc kubenswrapper[4966]: I1217 11:07:30.842803 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" path="/var/lib/kubelet/pods/e3fb435e-b61e-4aac-8629-9a07d685649e/volumes" Dec 17 11:08:46 crc kubenswrapper[4966]: I1217 11:08:46.807760 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 11:08:46 crc kubenswrapper[4966]: I1217 11:08:46.808336 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 11:09:16 crc kubenswrapper[4966]: I1217 11:09:16.808033 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 11:09:16 crc kubenswrapper[4966]: I1217 11:09:16.808460 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 11:09:46 crc kubenswrapper[4966]: I1217 11:09:46.807772 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 11:09:46 crc kubenswrapper[4966]: I1217 11:09:46.808769 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 11:09:46 crc kubenswrapper[4966]: I1217 11:09:46.808841 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 11:09:46 crc kubenswrapper[4966]: I1217 11:09:46.810522 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"774bf964c7159826b3f4e46131c91b45d09332fd4b4923df8be5ab7d71835abb"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 11:09:46 crc kubenswrapper[4966]: I1217 11:09:46.810610 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://774bf964c7159826b3f4e46131c91b45d09332fd4b4923df8be5ab7d71835abb" gracePeriod=600 Dec 17 11:09:46 crc kubenswrapper[4966]: I1217 11:09:46.945527 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="774bf964c7159826b3f4e46131c91b45d09332fd4b4923df8be5ab7d71835abb" exitCode=0 Dec 17 11:09:46 crc kubenswrapper[4966]: I1217 11:09:46.945586 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"774bf964c7159826b3f4e46131c91b45d09332fd4b4923df8be5ab7d71835abb"} Dec 17 11:09:46 crc kubenswrapper[4966]: I1217 11:09:46.945667 4966 scope.go:117] "RemoveContainer" containerID="97826eaaf06a692ee4bf2aac5700a37cd43c0b0119d810c8c8aba5175399905b" Dec 17 11:09:47 crc kubenswrapper[4966]: I1217 11:09:47.958612 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5"} Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.646906 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tm6hq"] Dec 17 11:10:54 crc kubenswrapper[4966]: E1217 11:10:54.652396 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerName="extract-content" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.652468 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerName="extract-content" Dec 17 11:10:54 crc kubenswrapper[4966]: E1217 11:10:54.652496 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="extract-utilities" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.652503 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="extract-utilities" Dec 17 11:10:54 crc kubenswrapper[4966]: E1217 11:10:54.652521 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="extract-content" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.652527 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="extract-content" Dec 17 11:10:54 crc kubenswrapper[4966]: E1217 11:10:54.652553 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="registry-server" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.652559 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="registry-server" Dec 17 11:10:54 crc kubenswrapper[4966]: E1217 11:10:54.652577 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerName="extract-utilities" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.652584 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerName="extract-utilities" Dec 17 11:10:54 crc kubenswrapper[4966]: E1217 11:10:54.652601 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerName="registry-server" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.652608 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerName="registry-server" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.653194 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fde326e-f683-4cc3-b90f-8d41f74541e6" containerName="registry-server" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.653328 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3fb435e-b61e-4aac-8629-9a07d685649e" containerName="registry-server" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.656060 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.681361 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm6hq"] Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.754729 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chxt8\" (UniqueName: \"kubernetes.io/projected/1d9681d7-583f-46f3-914b-aeb36583258a-kube-api-access-chxt8\") pod \"redhat-marketplace-tm6hq\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.754845 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-utilities\") pod \"redhat-marketplace-tm6hq\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.754957 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-catalog-content\") pod \"redhat-marketplace-tm6hq\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.856704 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chxt8\" (UniqueName: \"kubernetes.io/projected/1d9681d7-583f-46f3-914b-aeb36583258a-kube-api-access-chxt8\") pod \"redhat-marketplace-tm6hq\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.857141 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-utilities\") pod \"redhat-marketplace-tm6hq\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.857189 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-catalog-content\") pod \"redhat-marketplace-tm6hq\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.861240 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-catalog-content\") pod \"redhat-marketplace-tm6hq\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.872677 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-utilities\") pod \"redhat-marketplace-tm6hq\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:54 crc kubenswrapper[4966]: I1217 11:10:54.894677 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chxt8\" (UniqueName: \"kubernetes.io/projected/1d9681d7-583f-46f3-914b-aeb36583258a-kube-api-access-chxt8\") pod \"redhat-marketplace-tm6hq\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:55 crc kubenswrapper[4966]: I1217 11:10:55.037831 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:10:55 crc kubenswrapper[4966]: I1217 11:10:55.716131 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm6hq"] Dec 17 11:10:56 crc kubenswrapper[4966]: I1217 11:10:56.583567 4966 generic.go:334] "Generic (PLEG): container finished" podID="1d9681d7-583f-46f3-914b-aeb36583258a" containerID="4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334" exitCode=0 Dec 17 11:10:56 crc kubenswrapper[4966]: I1217 11:10:56.583666 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm6hq" event={"ID":"1d9681d7-583f-46f3-914b-aeb36583258a","Type":"ContainerDied","Data":"4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334"} Dec 17 11:10:56 crc kubenswrapper[4966]: I1217 11:10:56.583937 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm6hq" event={"ID":"1d9681d7-583f-46f3-914b-aeb36583258a","Type":"ContainerStarted","Data":"862796d89ec99f76e72b9ae62a0d00ae95c7d61fd57765f1a42f835ea1699018"} Dec 17 11:10:56 crc kubenswrapper[4966]: I1217 11:10:56.587138 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 11:10:58 crc kubenswrapper[4966]: I1217 11:10:58.602452 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm6hq" event={"ID":"1d9681d7-583f-46f3-914b-aeb36583258a","Type":"ContainerStarted","Data":"48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467"} Dec 17 11:11:00 crc kubenswrapper[4966]: I1217 11:11:00.624114 4966 generic.go:334] "Generic (PLEG): container finished" podID="1d9681d7-583f-46f3-914b-aeb36583258a" containerID="48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467" exitCode=0 Dec 17 11:11:00 crc kubenswrapper[4966]: I1217 11:11:00.624203 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm6hq" event={"ID":"1d9681d7-583f-46f3-914b-aeb36583258a","Type":"ContainerDied","Data":"48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467"} Dec 17 11:11:02 crc kubenswrapper[4966]: I1217 11:11:02.647959 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm6hq" event={"ID":"1d9681d7-583f-46f3-914b-aeb36583258a","Type":"ContainerStarted","Data":"6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca"} Dec 17 11:11:05 crc kubenswrapper[4966]: I1217 11:11:05.039361 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:11:05 crc kubenswrapper[4966]: I1217 11:11:05.040022 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:11:05 crc kubenswrapper[4966]: I1217 11:11:05.095724 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:11:05 crc kubenswrapper[4966]: I1217 11:11:05.124891 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tm6hq" podStartSLOduration=5.624536304 podStartE2EDuration="11.124359166s" podCreationTimestamp="2025-12-17 11:10:54 +0000 UTC" firstStartedPulling="2025-12-17 11:10:56.585600234 +0000 UTC m=+10192.130670176" lastFinishedPulling="2025-12-17 11:11:02.085423076 +0000 UTC m=+10197.630493038" observedRunningTime="2025-12-17 11:11:02.681340091 +0000 UTC m=+10198.226410053" watchObservedRunningTime="2025-12-17 11:11:05.124359166 +0000 UTC m=+10200.669429118" Dec 17 11:11:15 crc kubenswrapper[4966]: I1217 11:11:15.088058 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:11:15 crc kubenswrapper[4966]: I1217 11:11:15.145998 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm6hq"] Dec 17 11:11:15 crc kubenswrapper[4966]: I1217 11:11:15.769912 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tm6hq" podUID="1d9681d7-583f-46f3-914b-aeb36583258a" containerName="registry-server" containerID="cri-o://6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca" gracePeriod=2 Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.621459 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.780778 4966 generic.go:334] "Generic (PLEG): container finished" podID="1d9681d7-583f-46f3-914b-aeb36583258a" containerID="6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca" exitCode=0 Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.780818 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm6hq" event={"ID":"1d9681d7-583f-46f3-914b-aeb36583258a","Type":"ContainerDied","Data":"6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca"} Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.780850 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm6hq" event={"ID":"1d9681d7-583f-46f3-914b-aeb36583258a","Type":"ContainerDied","Data":"862796d89ec99f76e72b9ae62a0d00ae95c7d61fd57765f1a42f835ea1699018"} Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.780899 4966 scope.go:117] "RemoveContainer" containerID="6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.781269 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm6hq" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.783505 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-catalog-content\") pod \"1d9681d7-583f-46f3-914b-aeb36583258a\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.783568 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-utilities\") pod \"1d9681d7-583f-46f3-914b-aeb36583258a\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.783635 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chxt8\" (UniqueName: \"kubernetes.io/projected/1d9681d7-583f-46f3-914b-aeb36583258a-kube-api-access-chxt8\") pod \"1d9681d7-583f-46f3-914b-aeb36583258a\" (UID: \"1d9681d7-583f-46f3-914b-aeb36583258a\") " Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.784725 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-utilities" (OuterVolumeSpecName: "utilities") pod "1d9681d7-583f-46f3-914b-aeb36583258a" (UID: "1d9681d7-583f-46f3-914b-aeb36583258a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.798583 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d9681d7-583f-46f3-914b-aeb36583258a-kube-api-access-chxt8" (OuterVolumeSpecName: "kube-api-access-chxt8") pod "1d9681d7-583f-46f3-914b-aeb36583258a" (UID: "1d9681d7-583f-46f3-914b-aeb36583258a"). InnerVolumeSpecName "kube-api-access-chxt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.824023 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d9681d7-583f-46f3-914b-aeb36583258a" (UID: "1d9681d7-583f-46f3-914b-aeb36583258a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.827692 4966 scope.go:117] "RemoveContainer" containerID="48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.870502 4966 scope.go:117] "RemoveContainer" containerID="4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.886635 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.886697 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9681d7-583f-46f3-914b-aeb36583258a-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.886713 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chxt8\" (UniqueName: \"kubernetes.io/projected/1d9681d7-583f-46f3-914b-aeb36583258a-kube-api-access-chxt8\") on node \"crc\" DevicePath \"\"" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.906697 4966 scope.go:117] "RemoveContainer" containerID="6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca" Dec 17 11:11:16 crc kubenswrapper[4966]: E1217 11:11:16.908234 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca\": container with ID starting with 6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca not found: ID does not exist" containerID="6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.908388 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca"} err="failed to get container status \"6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca\": rpc error: code = NotFound desc = could not find container \"6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca\": container with ID starting with 6b5356fb8e43856f27a054585ab83b68f13b92ac294ab2ebe73a96f3d24817ca not found: ID does not exist" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.908430 4966 scope.go:117] "RemoveContainer" containerID="48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467" Dec 17 11:11:16 crc kubenswrapper[4966]: E1217 11:11:16.909025 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467\": container with ID starting with 48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467 not found: ID does not exist" containerID="48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.909088 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467"} err="failed to get container status \"48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467\": rpc error: code = NotFound desc = could not find container \"48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467\": container with ID starting with 48c00910a50ebbd0a7f46932f8dbb6128a2c0190ae812199ce70cdf156448467 not found: ID does not exist" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.909145 4966 scope.go:117] "RemoveContainer" containerID="4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334" Dec 17 11:11:16 crc kubenswrapper[4966]: E1217 11:11:16.911026 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334\": container with ID starting with 4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334 not found: ID does not exist" containerID="4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334" Dec 17 11:11:16 crc kubenswrapper[4966]: I1217 11:11:16.911059 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334"} err="failed to get container status \"4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334\": rpc error: code = NotFound desc = could not find container \"4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334\": container with ID starting with 4b9d17c81895fbe37d654484274f14f8bc7215e5bdc434e42feabe19005f6334 not found: ID does not exist" Dec 17 11:11:17 crc kubenswrapper[4966]: I1217 11:11:17.107177 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm6hq"] Dec 17 11:11:17 crc kubenswrapper[4966]: I1217 11:11:17.116544 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm6hq"] Dec 17 11:11:18 crc kubenswrapper[4966]: I1217 11:11:18.845529 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d9681d7-583f-46f3-914b-aeb36583258a" path="/var/lib/kubelet/pods/1d9681d7-583f-46f3-914b-aeb36583258a/volumes" Dec 17 11:12:16 crc kubenswrapper[4966]: I1217 11:12:16.809107 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 11:12:16 crc kubenswrapper[4966]: I1217 11:12:16.809684 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 11:12:46 crc kubenswrapper[4966]: I1217 11:12:46.810611 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 11:12:46 crc kubenswrapper[4966]: I1217 11:12:46.811334 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 11:13:16 crc kubenswrapper[4966]: I1217 11:13:16.808207 4966 patch_prober.go:28] interesting pod/machine-config-daemon-dxggm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 17 11:13:16 crc kubenswrapper[4966]: I1217 11:13:16.809717 4966 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 17 11:13:16 crc kubenswrapper[4966]: I1217 11:13:16.810087 4966 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" Dec 17 11:13:16 crc kubenswrapper[4966]: I1217 11:13:16.811161 4966 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5"} pod="openshift-machine-config-operator/machine-config-daemon-dxggm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 17 11:13:16 crc kubenswrapper[4966]: I1217 11:13:16.811411 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" containerName="machine-config-daemon" containerID="cri-o://845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" gracePeriod=600 Dec 17 11:13:16 crc kubenswrapper[4966]: E1217 11:13:16.941187 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:13:17 crc kubenswrapper[4966]: I1217 11:13:17.166840 4966 generic.go:334] "Generic (PLEG): container finished" podID="f703caec-d8d0-4e72-b58a-987e69356984" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" exitCode=0 Dec 17 11:13:17 crc kubenswrapper[4966]: I1217 11:13:17.166908 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerDied","Data":"845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5"} Dec 17 11:13:17 crc kubenswrapper[4966]: I1217 11:13:17.167403 4966 scope.go:117] "RemoveContainer" containerID="774bf964c7159826b3f4e46131c91b45d09332fd4b4923df8be5ab7d71835abb" Dec 17 11:13:17 crc kubenswrapper[4966]: I1217 11:13:17.167846 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:13:17 crc kubenswrapper[4966]: E1217 11:13:17.168512 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:13:30 crc kubenswrapper[4966]: I1217 11:13:30.830296 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:13:30 crc kubenswrapper[4966]: E1217 11:13:30.831580 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:13:44 crc kubenswrapper[4966]: I1217 11:13:44.837117 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:13:44 crc kubenswrapper[4966]: E1217 11:13:44.837959 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:13:59 crc kubenswrapper[4966]: I1217 11:13:59.830471 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:13:59 crc kubenswrapper[4966]: E1217 11:13:59.832199 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:14:12 crc kubenswrapper[4966]: I1217 11:14:12.830720 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:14:12 crc kubenswrapper[4966]: E1217 11:14:12.831564 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:14:25 crc kubenswrapper[4966]: I1217 11:14:25.831268 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:14:25 crc kubenswrapper[4966]: E1217 11:14:25.833717 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:14:40 crc kubenswrapper[4966]: I1217 11:14:40.831370 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:14:40 crc kubenswrapper[4966]: E1217 11:14:40.832377 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:14:53 crc kubenswrapper[4966]: I1217 11:14:53.832009 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:14:53 crc kubenswrapper[4966]: E1217 11:14:53.832694 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.441353 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg"] Dec 17 11:15:00 crc kubenswrapper[4966]: E1217 11:15:00.444751 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d9681d7-583f-46f3-914b-aeb36583258a" containerName="extract-content" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.444851 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d9681d7-583f-46f3-914b-aeb36583258a" containerName="extract-content" Dec 17 11:15:00 crc kubenswrapper[4966]: E1217 11:15:00.444940 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d9681d7-583f-46f3-914b-aeb36583258a" containerName="extract-utilities" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.444959 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d9681d7-583f-46f3-914b-aeb36583258a" containerName="extract-utilities" Dec 17 11:15:00 crc kubenswrapper[4966]: E1217 11:15:00.444982 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d9681d7-583f-46f3-914b-aeb36583258a" containerName="registry-server" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.444992 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d9681d7-583f-46f3-914b-aeb36583258a" containerName="registry-server" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.445944 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d9681d7-583f-46f3-914b-aeb36583258a" containerName="registry-server" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.449066 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.456561 4966 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.456711 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.478145 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg"] Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.585383 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lsrs\" (UniqueName: \"kubernetes.io/projected/9081d814-71d0-43c5-8287-9255cb99e545-kube-api-access-5lsrs\") pod \"collect-profiles-29432835-d7lzg\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.585485 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9081d814-71d0-43c5-8287-9255cb99e545-config-volume\") pod \"collect-profiles-29432835-d7lzg\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.586421 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9081d814-71d0-43c5-8287-9255cb99e545-secret-volume\") pod \"collect-profiles-29432835-d7lzg\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.699628 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lsrs\" (UniqueName: \"kubernetes.io/projected/9081d814-71d0-43c5-8287-9255cb99e545-kube-api-access-5lsrs\") pod \"collect-profiles-29432835-d7lzg\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.700027 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9081d814-71d0-43c5-8287-9255cb99e545-config-volume\") pod \"collect-profiles-29432835-d7lzg\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.700286 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9081d814-71d0-43c5-8287-9255cb99e545-secret-volume\") pod \"collect-profiles-29432835-d7lzg\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.706610 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9081d814-71d0-43c5-8287-9255cb99e545-config-volume\") pod \"collect-profiles-29432835-d7lzg\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.734336 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lsrs\" (UniqueName: \"kubernetes.io/projected/9081d814-71d0-43c5-8287-9255cb99e545-kube-api-access-5lsrs\") pod \"collect-profiles-29432835-d7lzg\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.740608 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9081d814-71d0-43c5-8287-9255cb99e545-secret-volume\") pod \"collect-profiles-29432835-d7lzg\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:00 crc kubenswrapper[4966]: I1217 11:15:00.791435 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:01 crc kubenswrapper[4966]: I1217 11:15:01.553299 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg"] Dec 17 11:15:02 crc kubenswrapper[4966]: I1217 11:15:02.178905 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" event={"ID":"9081d814-71d0-43c5-8287-9255cb99e545","Type":"ContainerStarted","Data":"fbfb8bc6661c79b81833973fc40097cf40d089b74c88c938b890e00ee4266657"} Dec 17 11:15:02 crc kubenswrapper[4966]: I1217 11:15:02.179343 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" event={"ID":"9081d814-71d0-43c5-8287-9255cb99e545","Type":"ContainerStarted","Data":"8debe253067ae4ed4711acf2d6c5d6cbc4a25056b266e3feff1bfe40fb2a9a9f"} Dec 17 11:15:02 crc kubenswrapper[4966]: I1217 11:15:02.206906 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" podStartSLOduration=2.205630404 podStartE2EDuration="2.205630404s" podCreationTimestamp="2025-12-17 11:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-17 11:15:02.201669127 +0000 UTC m=+10437.746739099" watchObservedRunningTime="2025-12-17 11:15:02.205630404 +0000 UTC m=+10437.750700346" Dec 17 11:15:03 crc kubenswrapper[4966]: I1217 11:15:03.193799 4966 generic.go:334] "Generic (PLEG): container finished" podID="9081d814-71d0-43c5-8287-9255cb99e545" containerID="fbfb8bc6661c79b81833973fc40097cf40d089b74c88c938b890e00ee4266657" exitCode=0 Dec 17 11:15:03 crc kubenswrapper[4966]: I1217 11:15:03.193921 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" event={"ID":"9081d814-71d0-43c5-8287-9255cb99e545","Type":"ContainerDied","Data":"fbfb8bc6661c79b81833973fc40097cf40d089b74c88c938b890e00ee4266657"} Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.207711 4966 generic.go:334] "Generic (PLEG): container finished" podID="efdaa7df-eeb5-48ec-afb5-a6c5cff26269" containerID="315a71f8fedc951dad75a270b142fbee024b98cfa6c046dcc1ced49081f4486d" exitCode=0 Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.207811 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"efdaa7df-eeb5-48ec-afb5-a6c5cff26269","Type":"ContainerDied","Data":"315a71f8fedc951dad75a270b142fbee024b98cfa6c046dcc1ced49081f4486d"} Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.564900 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.582024 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9081d814-71d0-43c5-8287-9255cb99e545-config-volume\") pod \"9081d814-71d0-43c5-8287-9255cb99e545\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.582419 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9081d814-71d0-43c5-8287-9255cb99e545-secret-volume\") pod \"9081d814-71d0-43c5-8287-9255cb99e545\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.583246 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9081d814-71d0-43c5-8287-9255cb99e545-config-volume" (OuterVolumeSpecName: "config-volume") pod "9081d814-71d0-43c5-8287-9255cb99e545" (UID: "9081d814-71d0-43c5-8287-9255cb99e545"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.583498 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lsrs\" (UniqueName: \"kubernetes.io/projected/9081d814-71d0-43c5-8287-9255cb99e545-kube-api-access-5lsrs\") pod \"9081d814-71d0-43c5-8287-9255cb99e545\" (UID: \"9081d814-71d0-43c5-8287-9255cb99e545\") " Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.584973 4966 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9081d814-71d0-43c5-8287-9255cb99e545-config-volume\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.592219 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9081d814-71d0-43c5-8287-9255cb99e545-kube-api-access-5lsrs" (OuterVolumeSpecName: "kube-api-access-5lsrs") pod "9081d814-71d0-43c5-8287-9255cb99e545" (UID: "9081d814-71d0-43c5-8287-9255cb99e545"). InnerVolumeSpecName "kube-api-access-5lsrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.592806 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9081d814-71d0-43c5-8287-9255cb99e545-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9081d814-71d0-43c5-8287-9255cb99e545" (UID: "9081d814-71d0-43c5-8287-9255cb99e545"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.686544 4966 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9081d814-71d0-43c5-8287-9255cb99e545-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.686588 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lsrs\" (UniqueName: \"kubernetes.io/projected/9081d814-71d0-43c5-8287-9255cb99e545-kube-api-access-5lsrs\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:04 crc kubenswrapper[4966]: I1217 11:15:04.842666 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:15:04 crc kubenswrapper[4966]: E1217 11:15:04.842934 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.220277 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" event={"ID":"9081d814-71d0-43c5-8287-9255cb99e545","Type":"ContainerDied","Data":"8debe253067ae4ed4711acf2d6c5d6cbc4a25056b266e3feff1bfe40fb2a9a9f"} Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.220322 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29432835-d7lzg" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.220526 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8debe253067ae4ed4711acf2d6c5d6cbc4a25056b266e3feff1bfe40fb2a9a9f" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.657224 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp"] Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.666199 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29432790-44tzp"] Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.687014 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.704800 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.704887 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config-secret\") pod \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.704959 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8445b\" (UniqueName: \"kubernetes.io/projected/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-kube-api-access-8445b\") pod \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.705040 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-workdir\") pod \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.705751 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ca-certs\") pod \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.705784 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-config-data\") pod \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.705810 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config\") pod \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.705860 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-temporary\") pod \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.705888 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ssh-key\") pod \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\" (UID: \"efdaa7df-eeb5-48ec-afb5-a6c5cff26269\") " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.706647 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-config-data" (OuterVolumeSpecName: "config-data") pod "efdaa7df-eeb5-48ec-afb5-a6c5cff26269" (UID: "efdaa7df-eeb5-48ec-afb5-a6c5cff26269"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.708004 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "efdaa7df-eeb5-48ec-afb5-a6c5cff26269" (UID: "efdaa7df-eeb5-48ec-afb5-a6c5cff26269"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.715291 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "efdaa7df-eeb5-48ec-afb5-a6c5cff26269" (UID: "efdaa7df-eeb5-48ec-afb5-a6c5cff26269"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.720555 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "efdaa7df-eeb5-48ec-afb5-a6c5cff26269" (UID: "efdaa7df-eeb5-48ec-afb5-a6c5cff26269"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.731821 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-kube-api-access-8445b" (OuterVolumeSpecName: "kube-api-access-8445b") pod "efdaa7df-eeb5-48ec-afb5-a6c5cff26269" (UID: "efdaa7df-eeb5-48ec-afb5-a6c5cff26269"). InnerVolumeSpecName "kube-api-access-8445b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.752930 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "efdaa7df-eeb5-48ec-afb5-a6c5cff26269" (UID: "efdaa7df-eeb5-48ec-afb5-a6c5cff26269"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.762778 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "efdaa7df-eeb5-48ec-afb5-a6c5cff26269" (UID: "efdaa7df-eeb5-48ec-afb5-a6c5cff26269"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.780003 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "efdaa7df-eeb5-48ec-afb5-a6c5cff26269" (UID: "efdaa7df-eeb5-48ec-afb5-a6c5cff26269"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.790611 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "efdaa7df-eeb5-48ec-afb5-a6c5cff26269" (UID: "efdaa7df-eeb5-48ec-afb5-a6c5cff26269"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.808380 4966 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.808514 4966 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.808532 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8445b\" (UniqueName: \"kubernetes.io/projected/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-kube-api-access-8445b\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.808545 4966 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.808558 4966 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.808572 4966 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-config-data\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.808584 4966 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.808597 4966 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.808609 4966 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdaa7df-eeb5-48ec-afb5-a6c5cff26269-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.831783 4966 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 17 11:15:05 crc kubenswrapper[4966]: I1217 11:15:05.910020 4966 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 17 11:15:06 crc kubenswrapper[4966]: I1217 11:15:06.229668 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"efdaa7df-eeb5-48ec-afb5-a6c5cff26269","Type":"ContainerDied","Data":"75f1df2d6e44e8ecea0dc9c0c2f7460793b15266299e63f02accb7db09e034fd"} Dec 17 11:15:06 crc kubenswrapper[4966]: I1217 11:15:06.230465 4966 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75f1df2d6e44e8ecea0dc9c0c2f7460793b15266299e63f02accb7db09e034fd" Dec 17 11:15:06 crc kubenswrapper[4966]: I1217 11:15:06.229728 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Dec 17 11:15:06 crc kubenswrapper[4966]: I1217 11:15:06.849186 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb2b6a0e-e006-4016-b0d7-5eca41e881b2" path="/var/lib/kubelet/pods/bb2b6a0e-e006-4016-b0d7-5eca41e881b2/volumes" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.722240 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 17 11:15:13 crc kubenswrapper[4966]: E1217 11:15:13.723437 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efdaa7df-eeb5-48ec-afb5-a6c5cff26269" containerName="tempest-tests-tempest-tests-runner" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.723454 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="efdaa7df-eeb5-48ec-afb5-a6c5cff26269" containerName="tempest-tests-tempest-tests-runner" Dec 17 11:15:13 crc kubenswrapper[4966]: E1217 11:15:13.723495 4966 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9081d814-71d0-43c5-8287-9255cb99e545" containerName="collect-profiles" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.723503 4966 state_mem.go:107] "Deleted CPUSet assignment" podUID="9081d814-71d0-43c5-8287-9255cb99e545" containerName="collect-profiles" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.723755 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="efdaa7df-eeb5-48ec-afb5-a6c5cff26269" containerName="tempest-tests-tempest-tests-runner" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.723787 4966 memory_manager.go:354] "RemoveStaleState removing state" podUID="9081d814-71d0-43c5-8287-9255cb99e545" containerName="collect-profiles" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.724588 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.734497 4966 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-j7vjx" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.782933 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.874744 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-778pd\" (UniqueName: \"kubernetes.io/projected/67f2d157-7930-4a7d-9a3a-8f116a1a5d6d-kube-api-access-778pd\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"67f2d157-7930-4a7d-9a3a-8f116a1a5d6d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.875035 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"67f2d157-7930-4a7d-9a3a-8f116a1a5d6d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.977166 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-778pd\" (UniqueName: \"kubernetes.io/projected/67f2d157-7930-4a7d-9a3a-8f116a1a5d6d-kube-api-access-778pd\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"67f2d157-7930-4a7d-9a3a-8f116a1a5d6d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.977668 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"67f2d157-7930-4a7d-9a3a-8f116a1a5d6d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 17 11:15:13 crc kubenswrapper[4966]: I1217 11:15:13.978346 4966 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"67f2d157-7930-4a7d-9a3a-8f116a1a5d6d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 17 11:15:14 crc kubenswrapper[4966]: I1217 11:15:14.002737 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-778pd\" (UniqueName: \"kubernetes.io/projected/67f2d157-7930-4a7d-9a3a-8f116a1a5d6d-kube-api-access-778pd\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"67f2d157-7930-4a7d-9a3a-8f116a1a5d6d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 17 11:15:14 crc kubenswrapper[4966]: I1217 11:15:14.015897 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"67f2d157-7930-4a7d-9a3a-8f116a1a5d6d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 17 11:15:14 crc kubenswrapper[4966]: I1217 11:15:14.095169 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 17 11:15:14 crc kubenswrapper[4966]: I1217 11:15:14.563739 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 17 11:15:15 crc kubenswrapper[4966]: I1217 11:15:15.319394 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"67f2d157-7930-4a7d-9a3a-8f116a1a5d6d","Type":"ContainerStarted","Data":"8ebd09d35244a81ac177a4b396c4596dbc7a80eee5d6cc26b81cd87f07e13e65"} Dec 17 11:15:16 crc kubenswrapper[4966]: I1217 11:15:16.328884 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"67f2d157-7930-4a7d-9a3a-8f116a1a5d6d","Type":"ContainerStarted","Data":"b23f931654c3520081c0209449c80b643841228b45fa0391287e097be84c34c3"} Dec 17 11:15:16 crc kubenswrapper[4966]: I1217 11:15:16.351198 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.995282257 podStartE2EDuration="3.351173113s" podCreationTimestamp="2025-12-17 11:15:13 +0000 UTC" firstStartedPulling="2025-12-17 11:15:14.570220759 +0000 UTC m=+10450.115290701" lastFinishedPulling="2025-12-17 11:15:15.926111615 +0000 UTC m=+10451.471181557" observedRunningTime="2025-12-17 11:15:16.340852013 +0000 UTC m=+10451.885921965" watchObservedRunningTime="2025-12-17 11:15:16.351173113 +0000 UTC m=+10451.896243085" Dec 17 11:15:19 crc kubenswrapper[4966]: I1217 11:15:19.831074 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:15:19 crc kubenswrapper[4966]: E1217 11:15:19.832474 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:15:34 crc kubenswrapper[4966]: I1217 11:15:34.838973 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:15:34 crc kubenswrapper[4966]: E1217 11:15:34.840359 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:15:45 crc kubenswrapper[4966]: I1217 11:15:45.831188 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:15:45 crc kubenswrapper[4966]: E1217 11:15:45.832010 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:15:56 crc kubenswrapper[4966]: I1217 11:15:56.831314 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:15:56 crc kubenswrapper[4966]: E1217 11:15:56.832628 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:16:06 crc kubenswrapper[4966]: I1217 11:16:06.872585 4966 scope.go:117] "RemoveContainer" containerID="56649688322a9ae67cfb6512b1c761a4cfb7a786d97f46f2654dad8a155752be" Dec 17 11:16:07 crc kubenswrapper[4966]: I1217 11:16:07.830859 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:16:07 crc kubenswrapper[4966]: E1217 11:16:07.832271 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:16:21 crc kubenswrapper[4966]: I1217 11:16:21.830615 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:16:21 crc kubenswrapper[4966]: E1217 11:16:21.832395 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:16:35 crc kubenswrapper[4966]: I1217 11:16:35.830978 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:16:35 crc kubenswrapper[4966]: E1217 11:16:35.831607 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:16:48 crc kubenswrapper[4966]: I1217 11:16:48.831301 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:16:48 crc kubenswrapper[4966]: E1217 11:16:48.832194 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:17:02 crc kubenswrapper[4966]: I1217 11:17:02.831411 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:17:02 crc kubenswrapper[4966]: E1217 11:17:02.832277 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:17:16 crc kubenswrapper[4966]: I1217 11:17:16.831753 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:17:16 crc kubenswrapper[4966]: E1217 11:17:16.833267 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:17:30 crc kubenswrapper[4966]: I1217 11:17:30.830861 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:17:30 crc kubenswrapper[4966]: E1217 11:17:30.833278 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:17:45 crc kubenswrapper[4966]: I1217 11:17:45.830472 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:17:45 crc kubenswrapper[4966]: E1217 11:17:45.831764 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:17:59 crc kubenswrapper[4966]: I1217 11:17:59.830193 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:17:59 crc kubenswrapper[4966]: E1217 11:17:59.830891 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.225043 4966 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xh8hd"] Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.228003 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.242194 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xh8hd"] Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.363689 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-utilities\") pod \"redhat-operators-xh8hd\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.363919 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-kube-api-access-tc8b2\") pod \"redhat-operators-xh8hd\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.363991 4966 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-catalog-content\") pod \"redhat-operators-xh8hd\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.466276 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-kube-api-access-tc8b2\") pod \"redhat-operators-xh8hd\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.466347 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-catalog-content\") pod \"redhat-operators-xh8hd\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.466426 4966 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-utilities\") pod \"redhat-operators-xh8hd\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.467241 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-utilities\") pod \"redhat-operators-xh8hd\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.467451 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-catalog-content\") pod \"redhat-operators-xh8hd\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.486823 4966 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-kube-api-access-tc8b2\") pod \"redhat-operators-xh8hd\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:06 crc kubenswrapper[4966]: I1217 11:18:06.552719 4966 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:07 crc kubenswrapper[4966]: I1217 11:18:07.067679 4966 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xh8hd"] Dec 17 11:18:08 crc kubenswrapper[4966]: I1217 11:18:08.035734 4966 generic.go:334] "Generic (PLEG): container finished" podID="5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" containerID="6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a" exitCode=0 Dec 17 11:18:08 crc kubenswrapper[4966]: I1217 11:18:08.035828 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xh8hd" event={"ID":"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0","Type":"ContainerDied","Data":"6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a"} Dec 17 11:18:08 crc kubenswrapper[4966]: I1217 11:18:08.035970 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xh8hd" event={"ID":"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0","Type":"ContainerStarted","Data":"36922af876cec8be8653658a48e86f6acf887dcca04305c12584f803abd83407"} Dec 17 11:18:08 crc kubenswrapper[4966]: I1217 11:18:08.040001 4966 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 17 11:18:10 crc kubenswrapper[4966]: I1217 11:18:10.055779 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xh8hd" event={"ID":"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0","Type":"ContainerStarted","Data":"8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8"} Dec 17 11:18:13 crc kubenswrapper[4966]: I1217 11:18:13.838233 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:18:13 crc kubenswrapper[4966]: E1217 11:18:13.840241 4966 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxggm_openshift-machine-config-operator(f703caec-d8d0-4e72-b58a-987e69356984)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" podUID="f703caec-d8d0-4e72-b58a-987e69356984" Dec 17 11:18:14 crc kubenswrapper[4966]: I1217 11:18:14.088850 4966 generic.go:334] "Generic (PLEG): container finished" podID="5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" containerID="8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8" exitCode=0 Dec 17 11:18:14 crc kubenswrapper[4966]: I1217 11:18:14.089024 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xh8hd" event={"ID":"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0","Type":"ContainerDied","Data":"8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8"} Dec 17 11:18:15 crc kubenswrapper[4966]: I1217 11:18:15.099532 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xh8hd" event={"ID":"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0","Type":"ContainerStarted","Data":"2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01"} Dec 17 11:18:15 crc kubenswrapper[4966]: I1217 11:18:15.122650 4966 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xh8hd" podStartSLOduration=2.591408223 podStartE2EDuration="9.122633873s" podCreationTimestamp="2025-12-17 11:18:06 +0000 UTC" firstStartedPulling="2025-12-17 11:18:08.039539663 +0000 UTC m=+10623.584609605" lastFinishedPulling="2025-12-17 11:18:14.570765313 +0000 UTC m=+10630.115835255" observedRunningTime="2025-12-17 11:18:15.117715749 +0000 UTC m=+10630.662785691" watchObservedRunningTime="2025-12-17 11:18:15.122633873 +0000 UTC m=+10630.667703815" Dec 17 11:18:16 crc kubenswrapper[4966]: I1217 11:18:16.553538 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:16 crc kubenswrapper[4966]: I1217 11:18:16.553910 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:17 crc kubenswrapper[4966]: I1217 11:18:17.613701 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xh8hd" podUID="5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" containerName="registry-server" probeResult="failure" output=< Dec 17 11:18:17 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 11:18:17 crc kubenswrapper[4966]: > Dec 17 11:18:24 crc kubenswrapper[4966]: I1217 11:18:24.838034 4966 scope.go:117] "RemoveContainer" containerID="845e2fffc60ca41bb51e93cd924adf01fde77155b9c836009d67fadac8b939c5" Dec 17 11:18:25 crc kubenswrapper[4966]: I1217 11:18:25.184926 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxggm" event={"ID":"f703caec-d8d0-4e72-b58a-987e69356984","Type":"ContainerStarted","Data":"2832079a872d4ab468b7cb7d06457c339402acfa695298ca57b32dfa479ac615"} Dec 17 11:18:27 crc kubenswrapper[4966]: I1217 11:18:27.618168 4966 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xh8hd" podUID="5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" containerName="registry-server" probeResult="failure" output=< Dec 17 11:18:27 crc kubenswrapper[4966]: timeout: failed to connect service ":50051" within 1s Dec 17 11:18:27 crc kubenswrapper[4966]: > Dec 17 11:18:36 crc kubenswrapper[4966]: I1217 11:18:36.610191 4966 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:36 crc kubenswrapper[4966]: I1217 11:18:36.683399 4966 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:37 crc kubenswrapper[4966]: I1217 11:18:37.419235 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xh8hd"] Dec 17 11:18:38 crc kubenswrapper[4966]: I1217 11:18:38.307073 4966 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xh8hd" podUID="5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" containerName="registry-server" containerID="cri-o://2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01" gracePeriod=2 Dec 17 11:18:38 crc kubenswrapper[4966]: I1217 11:18:38.894297 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.044037 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-utilities\") pod \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.044504 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-catalog-content\") pod \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.044706 4966 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-kube-api-access-tc8b2\") pod \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\" (UID: \"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0\") " Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.045174 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-utilities" (OuterVolumeSpecName: "utilities") pod "5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" (UID: "5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.045492 4966 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-utilities\") on node \"crc\" DevicePath \"\"" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.051183 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-kube-api-access-tc8b2" (OuterVolumeSpecName: "kube-api-access-tc8b2") pod "5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" (UID: "5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0"). InnerVolumeSpecName "kube-api-access-tc8b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.147680 4966 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-kube-api-access-tc8b2\") on node \"crc\" DevicePath \"\"" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.186624 4966 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" (UID: "5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.249048 4966 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.317155 4966 generic.go:334] "Generic (PLEG): container finished" podID="5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" containerID="2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01" exitCode=0 Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.317193 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xh8hd" event={"ID":"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0","Type":"ContainerDied","Data":"2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01"} Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.317219 4966 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xh8hd" event={"ID":"5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0","Type":"ContainerDied","Data":"36922af876cec8be8653658a48e86f6acf887dcca04305c12584f803abd83407"} Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.317237 4966 scope.go:117] "RemoveContainer" containerID="2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.317356 4966 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xh8hd" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.371940 4966 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xh8hd"] Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.399038 4966 scope.go:117] "RemoveContainer" containerID="8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8" Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.401416 4966 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xh8hd"] Dec 17 11:18:39 crc kubenswrapper[4966]: I1217 11:18:39.508335 4966 scope.go:117] "RemoveContainer" containerID="6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a" Dec 17 11:18:40 crc kubenswrapper[4966]: I1217 11:18:40.124788 4966 scope.go:117] "RemoveContainer" containerID="2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01" Dec 17 11:18:40 crc kubenswrapper[4966]: E1217 11:18:40.125782 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01\": container with ID starting with 2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01 not found: ID does not exist" containerID="2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01" Dec 17 11:18:40 crc kubenswrapper[4966]: I1217 11:18:40.125857 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01"} err="failed to get container status \"2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01\": rpc error: code = NotFound desc = could not find container \"2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01\": container with ID starting with 2ec90fdc9e89edc50efa130b97dcc1c52efc3178d75c37293942327ab97a4f01 not found: ID does not exist" Dec 17 11:18:40 crc kubenswrapper[4966]: I1217 11:18:40.125986 4966 scope.go:117] "RemoveContainer" containerID="8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8" Dec 17 11:18:40 crc kubenswrapper[4966]: E1217 11:18:40.126415 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8\": container with ID starting with 8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8 not found: ID does not exist" containerID="8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8" Dec 17 11:18:40 crc kubenswrapper[4966]: I1217 11:18:40.126514 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8"} err="failed to get container status \"8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8\": rpc error: code = NotFound desc = could not find container \"8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8\": container with ID starting with 8f84b97d73e2b99204817de4519df1211c7057da6e982e27d7ef4b6c0a0c50b8 not found: ID does not exist" Dec 17 11:18:40 crc kubenswrapper[4966]: I1217 11:18:40.126596 4966 scope.go:117] "RemoveContainer" containerID="6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a" Dec 17 11:18:40 crc kubenswrapper[4966]: E1217 11:18:40.126853 4966 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a\": container with ID starting with 6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a not found: ID does not exist" containerID="6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a" Dec 17 11:18:40 crc kubenswrapper[4966]: I1217 11:18:40.126971 4966 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a"} err="failed to get container status \"6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a\": rpc error: code = NotFound desc = could not find container \"6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a\": container with ID starting with 6e6d25110f09bb85fda2dc81341536d496b2a81672e793f561d49c2e20c0613a not found: ID does not exist" Dec 17 11:18:40 crc kubenswrapper[4966]: I1217 11:18:40.843466 4966 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0" path="/var/lib/kubelet/pods/5f5f7eb4-8270-4b6a-a398-b5ed3be27cc0/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515120510637024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015120510640017354 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015120463412016503 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015120463412015453 5ustar corecore